DocumentCode
2008373
Title
Multi-scale Q-learning of a mobile robot in dynamic environments
Author
Takase, Norio ; Kubota, Naoyuki ; Baba, N.
Author_Institution
Tokyo Metropolitan Univ., Tokyo, Japan
fYear
2012
fDate
20-24 Nov. 2012
Firstpage
1248
Lastpage
1252
Abstract
This paper deals with a state-dependent learning method of a mobile robot in dynamic and unknown environments. The aim of a mobile robot is to find the optimal path in the task of maze navigation on a grid world. Various types of reinforcement learning methods have been proposed, but it is very difficult to design the granularity (resolution) of states in search space. Therefore, we propose a multi-scale value function to enhance the initial learning of reinforcement learning. First, we compare the performance of temporal difference (TD) learning and Q-learning in dynamic environment. Here we assume several obstacles disappear in the grid world with an existence probability. Several experimental results show the effectiveness of the proposed method.
Keywords
collision avoidance; learning (artificial intelligence); mobile robots; probability; TD learning; existence probability; granularity design; grid world; maze navigation; mobile robot; multiscale Q-learning; multiscale value function; obstacle avoidance; reinforcement learning method; state-dependent learning method; temporal difference learning; Dynamic Environments; Intelligent Robitics; Multi-scale Value Function; Reinforcement Learning;
fLanguage
English
Publisher
ieee
Conference_Titel
Soft Computing and Intelligent Systems (SCIS) and 13th International Symposium on Advanced Intelligent Systems (ISIS), 2012 Joint 6th International Conference on
Conference_Location
Kobe
Print_ISBN
978-1-4673-2742-8
Type
conf
DOI
10.1109/SCIS-ISIS.2012.6505358
Filename
6505358
Link To Document