Title :
Memory-guided exploration in reinforcement learning
Author :
Carroll, James L. ; Peterson, Todd S. ; Owens, Nancy E.
Author_Institution :
Machine Intelligence Learning, & Decisions Lab., Brigham Young Univ., Provo, UT, USA
Abstract :
We focus on the task transfer in reinforcement learning and specifically in Q-learning. There are three main model free methods for performing task transfer in Q-learning: direct transfer, soft transfer and memory-guided exploration. In direct transfer, the Q-values from a previous task are used to initialize the Q-values of the next task. The soft transfer initializes the Q-values of the new task with a weighted average of the standard initialization value and the Q-values of the previous task. In memory-guided exploration the Q-values of previous tasks are used as a guide in the initial exploration of the agent. The weight that the agent gives to its past experience decreases over time. We explore stability issues related to the off-policy nature of memory-guided exploration and compare memory-guided exploration to soft transfer and direct transfer in three different environments
Keywords :
convergence; decision trees; learning (artificial intelligence); software agents; stability; Q-learning; convergence; decision trees; direct transfer; initialization value; machine learning; memory-guided exploration; reinforcement learning; soft transfer; stability; task transfer; Actuators; Contracts; Data engineering; Intelligent agent; Knowledge engineering; Laboratories; Machine intelligence; Machine learning; Robot sensing systems; Stability;
Conference_Titel :
Neural Networks, 2001. Proceedings. IJCNN '01. International Joint Conference on
Conference_Location :
Washington, DC
Print_ISBN :
0-7803-7044-9
DOI :
10.1109/IJCNN.2001.939497