• DocumentCode
    3400335
  • Title

    Multiagent reinforcement learning in extensive form games with complete information

  • Author

    Akramizadeh, Ali ; Menhaj, Mohammad-B ; Afshar, Ahmad

  • Author_Institution
    EE Dept., Polytech. Univ. of Tehran, Tehran
  • fYear
    2009
  • fDate
    March 30 2009-April 2 2009
  • Firstpage
    205
  • Lastpage
    211
  • Abstract
    Recent developments in multiagent reinforcement learning, mostly concentrate on normal form games or restrictive hierarchical form games. In this paper, we use the well known Q-learning in extensive form games which agents have a fixed priority in action selection. We also introduce a new concept called associative Q-values which not only can be used in action selection, leading to a subgame perfect equilibrium, but also can be used in update rule which is proved to be convergent. Associative Q-values are the expected utility of an agent in a game situation which is an estimate of the value of the subgame perfect equilibrium point.
  • Keywords
    game theory; learning (artificial intelligence); mathematics computing; multi-agent systems; Q-learning; associative Q-values; multiagent reinforcement learning; normal form games; restrictive hierarchical form games; subgame perfect equilibrium; Actuators; Collaboration; Distributed control; Game theory; Learning systems; Machine learning; Multiagent systems; Robot sensing systems; Sensor systems; Utility theory; Multiagent reinforcement learning; backward induction; exploration strategies; extensive form game; game theory; subgame perfect equilibrium;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Adaptive Dynamic Programming and Reinforcement Learning, 2009. ADPRL '09. IEEE Symposium on
  • Conference_Location
    Nashville, TN
  • Print_ISBN
    978-1-4244-2761-1
  • Type

    conf

  • DOI
    10.1109/ADPRL.2009.4927546
  • Filename
    4927546