Title :
Multiagent reinforcement learning in extensive form games with complete information
Author :
Akramizadeh, Ali ; Menhaj, Mohammad-B ; Afshar, Ahmad
Author_Institution :
EE Dept., Polytech. Univ. of Tehran, Tehran
fDate :
March 30 2009-April 2 2009
Abstract :
Recent developments in multiagent reinforcement learning, mostly concentrate on normal form games or restrictive hierarchical form games. In this paper, we use the well known Q-learning in extensive form games which agents have a fixed priority in action selection. We also introduce a new concept called associative Q-values which not only can be used in action selection, leading to a subgame perfect equilibrium, but also can be used in update rule which is proved to be convergent. Associative Q-values are the expected utility of an agent in a game situation which is an estimate of the value of the subgame perfect equilibrium point.
Keywords :
game theory; learning (artificial intelligence); mathematics computing; multi-agent systems; Q-learning; associative Q-values; multiagent reinforcement learning; normal form games; restrictive hierarchical form games; subgame perfect equilibrium; Actuators; Collaboration; Distributed control; Game theory; Learning systems; Machine learning; Multiagent systems; Robot sensing systems; Sensor systems; Utility theory; Multiagent reinforcement learning; backward induction; exploration strategies; extensive form game; game theory; subgame perfect equilibrium;
Conference_Titel :
Adaptive Dynamic Programming and Reinforcement Learning, 2009. ADPRL '09. IEEE Symposium on
Conference_Location :
Nashville, TN
Print_ISBN :
978-1-4244-2761-1
DOI :
10.1109/ADPRL.2009.4927546