DocumentCode :
3476050
Title :
Multiple reward criterion for cooperative behavior acquisition in a multiagent environment
Author :
Uchibe, Eiji ; Asada, Minoru
Author_Institution :
Dept. of Adaptive Machine Syst., Osaka Univ., Japan
Volume :
6
fYear :
1999
fDate :
1999
Firstpage :
710
Abstract :
A vector-valued reward function is discussed in the context of multiple behavior coordination, especially in a dynamically changing multiagent environment. Unlike the traditional weighted sum of several reward functions, we define a vector-valued value function which evaluates the current action strategy by introducing a discounted matrix to integrate several reward functions. Owing to the extension of the value function, the learning robot can estimate the future multiple reward from the environment appropriately not suffering from the weighting problem. The proposed method is applied to a simplified soccer game. Computer simulations are shown and a discussion is given
Keywords :
learning (artificial intelligence); mobile robots; multi-agent systems; multi-robot systems; singular value decomposition; cooperative behavior acquisition; discounted matrix; dynamically changing multiagent environment; learning robot; multiple reward criterion; soccer game; vector-valued reward function; vector-valued value function; Adaptive systems; Artificial intelligence; Decision making; Game theory; Learning; Machine vision; Orbital robotics; Predictive models; Robot kinematics; State-space methods;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Systems, Man, and Cybernetics, 1999. IEEE SMC '99 Conference Proceedings. 1999 IEEE International Conference on
Conference_Location :
Tokyo
ISSN :
1062-922X
Print_ISBN :
0-7803-5731-0
Type :
conf
DOI :
10.1109/ICSMC.1999.816638
Filename :
816638
Link To Document :
بازگشت