DocumentCode :
2718043
Title :
Convergence of multiagent Q-learning: Multi action replay process approach
Author :
Kim, Han-Eol ; Ahn, Hyo-Sung
Author_Institution :
Distrib. Control & Autonomous Syst. Lab., Gwangju Inst. of Sci. & Technol. (GIST), Gwangju, South Korea
fYear :
2010
fDate :
8-10 Sept. 2010
Firstpage :
789
Lastpage :
794
Abstract :
In this paper, we first suggest a new type of Markov model extended by Watkins´ action replay process. The new Markov model is called multi-action replay process (MARP), which is a process designed for multiagent coordination on the basis of reward values, state transition probabilities, and equilibrium strategy taking account of joint-action among agents. Using this model, multiagent Q-learning algorithm is then constructed as a cooperative reinforcement learning algorithm under completely connected agents. Finally, we prove that multiagent Q-learning values converge to optimal values. Simulation results are reported to illustrate the validity of the proposed multiagent Q-learning algorithm.
Keywords :
Markov processes; learning (artificial intelligence); multi-agent systems; MARP; Markov model; cooperative reinforcement learning algorithm; equilibrium strategy; multiaction replay process approach; multiagent Q-learning convergence; multiagent coordination; reward values; state transition probabilities; Algorithm design and analysis; Convergence; Equations; Games; Learning; Markov processes; Mathematical model;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Intelligent Control (ISIC), 2010 IEEE International Symposium on
Conference_Location :
Yokohama
ISSN :
2158-9860
Print_ISBN :
978-1-4244-5360-3
Electronic_ISBN :
2158-9860
Type :
conf
DOI :
10.1109/ISIC.2010.5612911
Filename :
5612911
Link To Document :
بازگشت