DocumentCode
3476050
Title
Multiple reward criterion for cooperative behavior acquisition in a multiagent environment
Author
Uchibe, Eiji ; Asada, Minoru
Author_Institution
Dept. of Adaptive Machine Syst., Osaka Univ., Japan
Volume
6
fYear
1999
fDate
1999
Firstpage
710
Abstract
A vector-valued reward function is discussed in the context of multiple behavior coordination, especially in a dynamically changing multiagent environment. Unlike the traditional weighted sum of several reward functions, we define a vector-valued value function which evaluates the current action strategy by introducing a discounted matrix to integrate several reward functions. Owing to the extension of the value function, the learning robot can estimate the future multiple reward from the environment appropriately not suffering from the weighting problem. The proposed method is applied to a simplified soccer game. Computer simulations are shown and a discussion is given
Keywords
learning (artificial intelligence); mobile robots; multi-agent systems; multi-robot systems; singular value decomposition; cooperative behavior acquisition; discounted matrix; dynamically changing multiagent environment; learning robot; multiple reward criterion; soccer game; vector-valued reward function; vector-valued value function; Adaptive systems; Artificial intelligence; Decision making; Game theory; Learning; Machine vision; Orbital robotics; Predictive models; Robot kinematics; State-space methods;
fLanguage
English
Publisher
ieee
Conference_Titel
Systems, Man, and Cybernetics, 1999. IEEE SMC '99 Conference Proceedings. 1999 IEEE International Conference on
Conference_Location
Tokyo
ISSN
1062-922X
Print_ISBN
0-7803-5731-0
Type
conf
DOI
10.1109/ICSMC.1999.816638
Filename
816638
Link To Document