DocumentCode :
3151040
Title :
Role differentiation process by division of reward function in multi-agent reinforcement learning
Author :
Taniguchi, Tadahiro ; Tabuchi, Kazuma ; Sawaragi, Tetsuo
Author_Institution :
Dept. of Human & Comput. Intell., Ritsumeikan Univ., Kusatsu
fYear :
2008
fDate :
20-22 Aug. 2008
Firstpage :
387
Lastpage :
393
Abstract :
There are several problems which discourage an organization from achieving tasks, e.g. partial observation, credit assignment, and concurrent learning in the multi-agent reinforcement learning domain. In many conventional approaches, each agent estimates hidden states, e.g., other agentspsila sensor inputs, positions, and policies, and reduces the uncertainty in the partially-observable Markov decision process (POMDP), and solves the multiagent reinforcement learning problem. In contrast, people reduce uncertainty in human organizations in the real world by autonommously dividing the roles played by each agent. In a framework of reinforcement learning, roles are mainly represented by goals for each agent. This paper presents a method for generating internal rewards from manager agents to worker agents. It also explicitly divides the roles, which can change a POMDP task for each agent into a simple MDP task under certain conditions. Several situational experiments are also described and the validity of the proposed method is evaluated.
Keywords :
Markov processes; learning (artificial intelligence); multi-agent systems; multiagent reinforcement learning; partially-observable Markov decision process; reward function; role differentiation; Animals; Competitive intelligence; Concurrent computing; Humans; Learning systems; Machine learning; Multiagent systems; Sensor systems; State estimation; Uncertainty; POMDP; multi-agent reinforcement learning; organization; reole differentiation;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
SICE Annual Conference, 2008
Conference_Location :
Tokyo
Print_ISBN :
978-4-907764-30-2
Electronic_ISBN :
978-4-907764-29-6
Type :
conf
DOI :
10.1109/SICE.2008.4654685
Filename :
4654685
Link To Document :
بازگشت