• DocumentCode
    2802102
  • Title

    Optimal rewards in multiagent teams

  • Author

    Bingyao Liu ; Singh, Sushil ; Lewis, Richard L. ; Qin, Shuang

  • Author_Institution
    Autom. Sci. & Engin., Beihang Univ., Beijing, China
  • fYear
    2012
  • fDate
    7-9 Nov. 2012
  • Firstpage
    1
  • Lastpage
    8
  • Abstract
    Following work on designing optimal rewards for single agents, we define a multiagent optimal rewards problem (ORP) in common-payoff (or team) settings. This new problem solves for individual agent reward functions that guide agents to better overall team performance relative to teams in which all agents guide their behavior with the same given team-reward function. We present a multiagent architecture in which each agent learns good reward functions from experience using a gradient-based algorithm in addition to performing the usual task of planning good policies (except in this case with respect to the learned rather than the given reward function). Multiagency introduces the challenge of nonstationarity: because the agents learn simultaneously, each agent´s learning problem is nonstationary and interdependent on the other agents. We demonstrate on two simple domains that the proposed architecture outperforms the conventional approach in which all the agents use the same given team-reward function (even when accounting for the resource overhead of the reward learning); that the learning algorithm performs stably despite the nonstationarity; and that learning individual reward functions can lead to better specialization of roles than is possible with shared reward, whether learned or given.
  • Keywords
    learning (artificial intelligence); multi-agent systems; optimisation; planning (artificial intelligence); agent behavior; agent learning problem; common-payoff setting; decentralized planning approach; gradient-based algorithm; individual agent reward function; learning algorithm; multiagent ORP; multiagent architecture; multiagent optimal rewards problem; multiagent team; nonstationarity; optimization approach; resource overhead; reward learning; role specialization; shared reward; team performance; team-reward function; Algorithm design and analysis; Computer architecture; History; Joints; Learning systems; Planning; Search problems;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Development and Learning and Epigenetic Robotics (ICDL), 2012 IEEE International Conference on
  • Conference_Location
    San Diego, CA
  • Print_ISBN
    978-1-4673-4964-2
  • Electronic_ISBN
    978-1-4673-4963-5
  • Type

    conf

  • DOI
    10.1109/DevLrn.2012.6400862
  • Filename
    6400862