DocumentCode
3400335
Title
Multiagent reinforcement learning in extensive form games with complete information
Author
Akramizadeh, Ali ; Menhaj, Mohammad-B ; Afshar, Ahmad
Author_Institution
EE Dept., Polytech. Univ. of Tehran, Tehran
fYear
2009
fDate
March 30 2009-April 2 2009
Firstpage
205
Lastpage
211
Abstract
Recent developments in multiagent reinforcement learning, mostly concentrate on normal form games or restrictive hierarchical form games. In this paper, we use the well known Q-learning in extensive form games which agents have a fixed priority in action selection. We also introduce a new concept called associative Q-values which not only can be used in action selection, leading to a subgame perfect equilibrium, but also can be used in update rule which is proved to be convergent. Associative Q-values are the expected utility of an agent in a game situation which is an estimate of the value of the subgame perfect equilibrium point.
Keywords
game theory; learning (artificial intelligence); mathematics computing; multi-agent systems; Q-learning; associative Q-values; multiagent reinforcement learning; normal form games; restrictive hierarchical form games; subgame perfect equilibrium; Actuators; Collaboration; Distributed control; Game theory; Learning systems; Machine learning; Multiagent systems; Robot sensing systems; Sensor systems; Utility theory; Multiagent reinforcement learning; backward induction; exploration strategies; extensive form game; game theory; subgame perfect equilibrium;
fLanguage
English
Publisher
ieee
Conference_Titel
Adaptive Dynamic Programming and Reinforcement Learning, 2009. ADPRL '09. IEEE Symposium on
Conference_Location
Nashville, TN
Print_ISBN
978-1-4244-2761-1
Type
conf
DOI
10.1109/ADPRL.2009.4927546
Filename
4927546
Link To Document