Title :
Q-learning algorithm using an adaptive-sized Q-table
Author :
Hirashima, Yoichi ; Iiguni, Youji ; Inoue, Akira ; Masuda, Shiro
Author_Institution :
Dept. of Syst. Eng., Okayama Univ., Japan
fDate :
6/21/1905 12:00:00 AM
Abstract :
Q-learning is one of the successfully established algorithms for the reinforcement learning. It directly estimates optimal Q-values for pairs of states and admissible actions. By using these Q-values, agents can obtain the optimal movement in controlled Markovian domains without using an explicit model of the desired system. However, the algorithm requires a large number of actions of trial and error in the early stages of learning. In this paper, a Q-learning algorithm using the Memory Based Learning (MBL) system is proposed. By using the generalization property of the MBL system, the learning effect for a Q-value can be spread to adjacent Q-values, and thus the number of actions of trial and error can be reduced. Finally, computer simulation results for the control of inverted pendulums are presented to show the effectiveness of the proposed method
Keywords :
Markov processes; learning (artificial intelligence); stability; Q-learning algorithm; adaptive-sized Q-table; admissible actions; computer simulation; controlled Markovian domains; inverted pendulums; memory based learning; reinforcement learning; Cities and towns; Ear; Educational institutions; Humans; Learning; Modeling; Optimal control; Quantization; Systems engineering and theory; Table lookup;
Conference_Titel :
Decision and Control, 1999. Proceedings of the 38th IEEE Conference on
Conference_Location :
Phoenix, AZ
Print_ISBN :
0-7803-5250-5
DOI :
10.1109/CDC.1999.830250