Title :
Multi-agent learning methods in an uncertain environment
Author :
Liu, Shu-hua ; Tian, Yan-tao
Abstract :
In this paper, the multi-agent learning methods in an uncertain environment are addressed. The advantages and disadvantages of each algorithm are given. Rationality and convergence are the two main properties of multi-agent learning algorithms. However, it is very difficult to achieve both properties simultaneously. Minmax-Q learning is guaranteed to converge to equilibrium but there is no guarantee that this is the best response to the actual opponent. Therefore, Minmax-Q is not rational. In contrast, opponent modeling is rational but not convergent. Reinforcement learning using a variable learning rate and simultaneously achieves both properties. To reduce the dimension of state space, modular Q-learning and multilayered reinforcement learning are presented. The presented methods are not exhaustive, but they highlight the major methods used by researchers in the past years.
Keywords :
Markov processes; convergence; game theory; learning (artificial intelligence); multi-agent systems; MDP; Markov decision process problems; altruistic reinforcement learning; artificial intelligence; convergence; minimax-Q learning; multi-agent learning methods; opponent modeling; profit-sharing reinforcement learning; rationality; reinforcement learning; team-partitioned opaque-transition reinforcement learning; uncertain environment; variable learning rate; Application software; Artificial intelligence; Computer science; Learning systems; Machine learning; Multiagent systems; Output feedback; Problem-solving; Supervised learning; Unsupervised learning;
Conference_Titel :
Machine Learning and Cybernetics, 2002. Proceedings. 2002 International Conference on
Print_ISBN :
0-7803-7508-4
DOI :
10.1109/ICMLC.2002.1174416