DocumentCode :
3110531
Title :
Accelerated Q-learning for fail state and action spaces
Author :
Park, In-Won ; Kim, Jong-Hwan ; Park, Kui-Hong
Author_Institution :
Dept. of Electr. Eng. & Comput. Sci., KAIST, Daejeon
fYear :
2008
fDate :
12-15 Oct. 2008
Firstpage :
763
Lastpage :
767
Abstract :
Accelerated Q-learning algorithm is proposed for environment having both goal and fail states. It extends Q-learning, a well-known scheme in reinforcement learning. Unlike this conventional Q-learning, the proposed algorithm keeps track of the past failure experiences as a separate fail state-action value, QF. Agent uses this value along with a goal state-action value, QN, which is calculated and updated using conventional Q-learning, to modify the exploratory behavior during learning phase. Effectiveness of the proposed accelerated Q-learning algorithm is verified in a grid world environment. The proposed algorithm significantly reduces a convergence speed to find out the optimal path from start state to goal state while maximizing its receiving rewards.
Keywords :
grid computing; learning (artificial intelligence); accelerated Q-learning; fail state spaces; grid world environment; reinforcement learning; Acceleration; Computer science; Convergence; Humanoid robots; Humans; Machine learning algorithms; Negative feedback; Space technology; Telecommunications; Thin film transistors; Accelerated Q-learning algorithm; success and failure experiences of the agent;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Systems, Man and Cybernetics, 2008. SMC 2008. IEEE International Conference on
Conference_Location :
Singapore
ISSN :
1062-922X
Print_ISBN :
978-1-4244-2383-5
Electronic_ISBN :
1062-922X
Type :
conf
DOI :
10.1109/ICSMC.2008.4811370
Filename :
4811370
Link To Document :
بازگشت