Title :
Improving the efficiency of Bayesian inverse reinforcement learning
Author :
Michini, Bernard ; How, Jonathan P.
Author_Institution :
Aerosp. Controls Lab., Massachusetts Inst. of Technol., Cambridge, MA, USA
Abstract :
Inverse reinforcement learning (IRL) is the task of learning the reward function of a Markov Decision Process (MDP) given knowledge of the transition function and a set of expert demonstrations. While many IRL algorithms exist, Bayesian IRL [1] provides a general and principled method of reward learning by casting the problem in the Bayesian inference framework. However, the algorithm as originally presented suffers from several inefficiencies that prohibit its use for even moderate problem sizes. This paper proposes modifications to the original Bayesian IRL algorithm to improve its efficiency and tractability in situations where the state space is large and the expert demonstrations span only a small portion of it. The key insight is that the inference task should be focused on states that are similar to those encountered by the expert, as opposed to making the naive assumption that the expert demonstrations contain enough information to accurately infer the reward function over the entire state space. A modified algorithm is presented and experimental results show substantially faster convergence while maintaining the solution quality of the original method.
Keywords :
Markov processes; belief networks; inference mechanisms; learning (artificial intelligence); Bayesian inference framework; Bayesian inverse reinforcement learning; IRL algorithms; Markov decision process; convergence; expert demonstrations; inference task; transition function; Bayesian methods; Cooling; Inference algorithms; Kernel; Schedules; Standards; Vectors;
Conference_Titel :
Robotics and Automation (ICRA), 2012 IEEE International Conference on
Conference_Location :
Saint Paul, MN
Print_ISBN :
978-1-4673-1403-9
Electronic_ISBN :
1050-4729
DOI :
10.1109/ICRA.2012.6225241