Title :
Learning task decomposition and exploration shaping for reinforcement learning agents
Author :
Djurdjevic, Predrag ; Huber, Manfred
Author_Institution :
Dept. of Comput. Sci. & Eng., Univ. of Texas at Arlington, Arlington, TX
Abstract :
For situated reinforcement learning agents to succeed in complex real world environments they have to be able to efficiently acquire and reuse control knowledge in order to accomplish new tasks faster and to accelerate the learning of new policies. While hierarchical learning approaches which transfer previously acquired skills and representations to model and control new tasks have the potential to significantly improve learning times, they also pose the risk of ldquobehavior proliferationrdquo where the growing set of available actions makes it increasingly difficult to determine a strategy for a new task. To overcome this problem and to further improve knowledge reuse, the learning agent should thus also have the ability to predict the utility of an action or reusable skill in a new context and to analyze new tasks in order to decompose them into known subtasks. This paper presents a novel approach for learning task decomposition by learning to predict the utility of subgoals and subgoal types in the context of a new task, as well as for exploration shaping by predicting the likelihood with which each available action is useful in the given task context. This information, encoded as a set of utility functions, is then used to focus the exploration and learning process of the agent to increase performance both in terms of the time spent to reach the new task´s goal the first time and of the time required to learn an optimal policy. This ability is demonstrated here in the context of navigation and manipulation tasks in a feature enhanced grid world domain.
Keywords :
knowledge representation; learning (artificial intelligence); multi-agent systems; behavior proliferation; complex real world environment; exploration shaping; feature enhanced grid world domain; knowledge representation; knowledge reuse; manipulation task; navigation task; reinforcement learning agent; task decomposition; utility function; Accelerated aging; Acceleration; Computer science; Data mining; Knowledge engineering; Learning; Monte Carlo methods; Navigation; Shape control; State-space methods; exploration shaping; reinforcement learning; task decomposition; transfer learning;
Conference_Titel :
Systems, Man and Cybernetics, 2008. SMC 2008. IEEE International Conference on
Conference_Location :
Singapore
Print_ISBN :
978-1-4244-2383-5
Electronic_ISBN :
1062-922X
DOI :
10.1109/ICSMC.2008.4811303