Learning Transition Dynamics via Rewarded Exploration: A Study using Unity's MLAgents

Public Deposited
Resource Type
Creator
Abstract
  • AI agents can benefit from understanding their environment and how it works, as being able to predict the state of the environment after one makes an action is useful for doing tasks. My work explores using a custom reward system to guide an AI agent in learning the transition dynamics of its environment via exploration. Due to the popularity of game engines, I focus on building a transition dynamics model using the game engine, Unity, which provides a package for making AI agents. I test the agent's behaviour across 8 studies, with different hyperparameters for its neural network and with and without access to memory via Long Short-Term Memory. I also conducted two tests with a different reward system to help judge the effectiveness of my approach. The results of my experiments show that the agent performs well and is capable of predicting a variable in the environment.

Subject
Language
Publisher
Thesis Degree Level
Thesis Degree Name
Thesis Degree Discipline
Identifier
Rights Notes
  • Copyright © 2022 the author(s). Theses may be used for non-commercial research, educational, or related academic purposes only. Such uses include personal study, research, scholarship, and teaching. Theses may only be shared by linking to Carleton University Institutional Repository and no part may be used without proper attribution to the author. No part may be used for commercial purposes directly or indirectly via a for-profit platform; no adaptation or derivative works are permitted without consent from the copyright owner.

Date Created
  • 2022

Relations

In Collection:

Items