• DocumentCode
    3476050
  • Title

    Multiple reward criterion for cooperative behavior acquisition in a multiagent environment

  • Author

    Uchibe, Eiji ; Asada, Minoru

  • Author_Institution
    Dept. of Adaptive Machine Syst., Osaka Univ., Japan
  • Volume
    6
  • fYear
    1999
  • fDate
    1999
  • Firstpage
    710
  • Abstract
    A vector-valued reward function is discussed in the context of multiple behavior coordination, especially in a dynamically changing multiagent environment. Unlike the traditional weighted sum of several reward functions, we define a vector-valued value function which evaluates the current action strategy by introducing a discounted matrix to integrate several reward functions. Owing to the extension of the value function, the learning robot can estimate the future multiple reward from the environment appropriately not suffering from the weighting problem. The proposed method is applied to a simplified soccer game. Computer simulations are shown and a discussion is given
  • Keywords
    learning (artificial intelligence); mobile robots; multi-agent systems; multi-robot systems; singular value decomposition; cooperative behavior acquisition; discounted matrix; dynamically changing multiagent environment; learning robot; multiple reward criterion; soccer game; vector-valued reward function; vector-valued value function; Adaptive systems; Artificial intelligence; Decision making; Game theory; Learning; Machine vision; Orbital robotics; Predictive models; Robot kinematics; State-space methods;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Systems, Man, and Cybernetics, 1999. IEEE SMC '99 Conference Proceedings. 1999 IEEE International Conference on
  • Conference_Location
    Tokyo
  • ISSN
    1062-922X
  • Print_ISBN
    0-7803-5731-0
  • Type

    conf

  • DOI
    10.1109/ICSMC.1999.816638
  • Filename
    816638