DocumentCode :
2568631
Title :
A reinforcement learning model using macro-actions in multi-task grid-world problems
Author :
Onda, Hiroshi ; Ozawa, Seiichi
Author_Institution :
Grad. Sch. of Eng., Kobe Univ., Kobe, Japan
fYear :
2009
fDate :
11-14 Oct. 2009
Firstpage :
3088
Lastpage :
3093
Abstract :
A macro-action is a typical series of useful actions that brings high expected rewards to an agent. Murata et al. have proposed an actor-critic model which can generate macro-actions automatically based on the information on state values and visiting frequency of states. However, their model has not assumed that generated macro-actions are utilized for leaning different tasks. In this paper, we extend the Murata´s model such that generated macro-actions can help an agent learn an optimal policy quickly in multi-task grid-world (MTGW) maze problems. The proposed model is applied to two MTGW problems, each of which consists of six different maze tasks. From the experimental results, it is concluded that the proposed model could speed up learning if macro-actions are generated in the so-called correlated regions.
Keywords :
grid computing; learning (artificial intelligence); multi-agent systems; neural nets; MTGW maze problem; Murata´s model; actor-critic model; agent learning; macro-action; multitask grid-world problem; neural network; reinforcement learning model; Automatic control; Cybernetics; Frequency; Humans; Learning; Mesh generation; Neural networks; Resource management; Temperature; USA Councils; macro-action; multitask learning; neural network; reinforcement learning;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Systems, Man and Cybernetics, 2009. SMC 2009. IEEE International Conference on
Conference_Location :
San Antonio, TX
ISSN :
1062-922X
Print_ISBN :
978-1-4244-2793-2
Electronic_ISBN :
1062-922X
Type :
conf
DOI :
10.1109/ICSMC.2009.5346139
Filename :
5346139
Link To Document :
بازگشت