DocumentCode
3087796
Title
Effect of human guidance and state space size on Interactive Reinforcement Learning
Author
Suay, Halit Bener ; Chernova, Sonia
Author_Institution
Robot. Eng. Program, Worcester Polytech. Inst., Worcester, MA, USA
fYear
2011
fDate
July 31 2011-Aug. 3 2011
Firstpage
1
Lastpage
6
Abstract
The Interactive Reinforcement Learning algorithm enables a human user to train a robot by providing rewards in response to past actions and anticipatory guidance to guide the selection of future actions. Past work with software agents has shown that incorporating user guidance into the policy learning process through Interactive Reinforcement Learning significantly improves the policy learning time by reducing the number of states the agent explores. We present the first study of Interactive Reinforcement Learning in real-world robotic systems. We report on four experiments that study the effects that teacher guidance and state space size have on policy learning performance. We discuss modifications made to apply Interactive Reinforcement Learning to a real-world system and show that guidance significantly reduces the learning rate, and that its positive effects increase with state space size.
Keywords
human-robot interaction; interactive systems; learning (artificial intelligence); human guidance; interactive reinforcement learning; policy learning process; real-world robotic systems; software agents; state space size; Entropy; Humans; Learning; Robots; Strontium; Training;
fLanguage
English
Publisher
ieee
Conference_Titel
RO-MAN, 2011 IEEE
Conference_Location
Atlanta, GA
Print_ISBN
978-1-4577-1571-6
Electronic_ISBN
978-1-4577-1572-3
Type
conf
DOI
10.1109/ROMAN.2011.6005223
Filename
6005223
Link To Document