Title :
Role differentiation process by division of reward function in multi-agent reinforcement learning
Author :
Taniguchi, Tadahiro ; Tabuchi, Kazuma ; Sawaragi, Tetsuo
Author_Institution :
Dept. of Human & Comput. Intell., Ritsumeikan Univ., Kusatsu
Abstract :
There are several problems which discourage an organization from achieving tasks, e.g. partial observation, credit assignment, and concurrent learning in the multi-agent reinforcement learning domain. In many conventional approaches, each agent estimates hidden states, e.g., other agentspsila sensor inputs, positions, and policies, and reduces the uncertainty in the partially-observable Markov decision process (POMDP), and solves the multiagent reinforcement learning problem. In contrast, people reduce uncertainty in human organizations in the real world by autonommously dividing the roles played by each agent. In a framework of reinforcement learning, roles are mainly represented by goals for each agent. This paper presents a method for generating internal rewards from manager agents to worker agents. It also explicitly divides the roles, which can change a POMDP task for each agent into a simple MDP task under certain conditions. Several situational experiments are also described and the validity of the proposed method is evaluated.
Keywords :
Markov processes; learning (artificial intelligence); multi-agent systems; multiagent reinforcement learning; partially-observable Markov decision process; reward function; role differentiation; Animals; Competitive intelligence; Concurrent computing; Humans; Learning systems; Machine learning; Multiagent systems; Sensor systems; State estimation; Uncertainty; POMDP; multi-agent reinforcement learning; organization; reole differentiation;
Conference_Titel :
SICE Annual Conference, 2008
Conference_Location :
Tokyo
Print_ISBN :
978-4-907764-30-2
Electronic_ISBN :
978-4-907764-29-6
DOI :
10.1109/SICE.2008.4654685