Title :
Apprenticeship learning via soft local homomorphisms
Author :
Boularias, Abdeslam ; Chaib-Draa, Brahim
Author_Institution :
Comput. Sci. & Software Eng. Dept., Laval Univ., Quebec City, QC, Canada
Abstract :
We consider the problem of apprenticeship learning when the expert´s demonstration covers only a small part of a large state space. Inverse Reinforcement Learning (IRL) provides an efficient solution to this problem based on the assumption that the expert is optimally acting in a Markov Decision Process (MDP). However, past work on IRL requires an accurate estimate of the frequency of encountering each feature of the states when the robot follows the expert´s policy. Given that the complete policy of the expert is unknown, the features frequencies can only be empirically estimated from the demonstrated trajectories. In this paper, we propose to use a transfer method, known as soft homomorphism, in order to generalize the expert´s policy to unvisited regions of the state space. The generalized policy can be used either as the robot´s final policy, or to calculate the features frequencies within an IRL algorithm. Empirical results show that our approach is able to learn good policies from a small number of demonstrations.
Keywords :
Markov processes; learning (artificial intelligence); robots; state-space methods; Markov decision process; apprenticeship learning; features frequency; inverse reinforcement learning; robot final policy; soft local homomorphism; state space; transfer method; Computer science; Frequency estimation; Learning; Orbital robotics; Robotics and automation; Robots; Software engineering; State estimation; State-space methods; USA Councils;
Conference_Titel :
Robotics and Automation (ICRA), 2010 IEEE International Conference on
Conference_Location :
Anchorage, AK
Print_ISBN :
978-1-4244-5038-1
Electronic_ISBN :
1050-4729
DOI :
10.1109/ROBOT.2010.5509717