DocumentCode :
3324449
Title :
Memory-guided exploration in reinforcement learning
Author :
Carroll, James L. ; Peterson, Todd S. ; Owens, Nancy E.
Author_Institution :
Machine Intelligence Learning, & Decisions Lab., Brigham Young Univ., Provo, UT, USA
Volume :
2
fYear :
2001
fDate :
2001
Firstpage :
1002
Abstract :
We focus on the task transfer in reinforcement learning and specifically in Q-learning. There are three main model free methods for performing task transfer in Q-learning: direct transfer, soft transfer and memory-guided exploration. In direct transfer, the Q-values from a previous task are used to initialize the Q-values of the next task. The soft transfer initializes the Q-values of the new task with a weighted average of the standard initialization value and the Q-values of the previous task. In memory-guided exploration the Q-values of previous tasks are used as a guide in the initial exploration of the agent. The weight that the agent gives to its past experience decreases over time. We explore stability issues related to the off-policy nature of memory-guided exploration and compare memory-guided exploration to soft transfer and direct transfer in three different environments
Keywords :
convergence; decision trees; learning (artificial intelligence); software agents; stability; Q-learning; convergence; decision trees; direct transfer; initialization value; machine learning; memory-guided exploration; reinforcement learning; soft transfer; stability; task transfer; Actuators; Contracts; Data engineering; Intelligent agent; Knowledge engineering; Laboratories; Machine intelligence; Machine learning; Robot sensing systems; Stability;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2001. Proceedings. IJCNN '01. International Joint Conference on
Conference_Location :
Washington, DC
ISSN :
1098-7576
Print_ISBN :
0-7803-7044-9
Type :
conf
DOI :
10.1109/IJCNN.2001.939497
Filename :
939497
Link To Document :
بازگشت