Title :
Learn to Coordinate with Generic Non-Stationary Opponents
Author_Institution :
Dept. of Comput. Sci. & Technol., Tsinghua Univ., Beijing
Abstract :
Learning to coordinate with non-stationary opponents is a major challenge for adaptive agents. Most previous research investigated only restricted classes of such dynamic opponents. The main contribution of this paper is twofold: (i) A class of generic non-stationary opponents is introduced. The opponents keep mixed strategies which change with less regularity. Its showed that the independent reinforcement learners (ILs), which have neither prior knowledge nor opponent models, cannot coordinate well with this type of opponent, (ii) A new exploration strategy, the DAE (detect and explore) mechanism, is tailored for the ILs in such coordination tasks. This mechanism allows the ILs dynamically detect changes in the opponents behavior and adjust their learning rate and exploration temperature. It´s showed that ILs using this strategy are still able to converge in self-play and are able to coordinate well with the non-stationary opponents
Keywords :
learning (artificial intelligence); multi-agent systems; adaptive agents; detect and explore mechanism; dynamic opponents; exploration strategy; generic nonstationary opponents; independent reinforcement learner; Cognitive informatics; Computer science; Decision making; Game theory; Learning; Multiagent systems; Temperature; Testing; Coordination game; Exploration strategy; Non-stationary opponent; Reinforcement learning;
Conference_Titel :
Cognitive Informatics, 2006. ICCI 2006. 5th IEEE International Conference on
Conference_Location :
Beijing
Print_ISBN :
1-4244-0475-4
DOI :
10.1109/COGINF.2006.365546