DocumentCode
3410074
Title
Trajectory-model-based reinforcement learning: Application to bimanual humanoid motor learning with a closed-chain constraint
Author
Sugimoto, Norikazu ; Morimoto, Jun
Author_Institution
Dept. of Brain Machine Interface, Nat. Inst. of Inf. & Commun. Technol., Suita, Japan
fYear
2013
fDate
15-17 Oct. 2013
Firstpage
429
Lastpage
434
Abstract
We propose a reinforcement learning (RL) framework to improve policies for a high-dimensional system through fewer interactions with real environments than standard RL methods. In our learning framework, we first use off-line simulations to improve the controller parameters with an approximated environment model to generate samples along locally optimized trajectories. We then use the approximated dynamics to improve the performance of a tool manipulation task in a path integral RL framework, which updates a policy from the sampled trajectories of the state and action vectors and the cost. In this study, we apply our proposed method to a bimanual humanoid motor learning task in which we need to explicitly consider a closed-chain constraint. We show that a 51-DOF real humanoid robot can learn to manipulate a rod to hit via-points using both arms within 36 interactions in a real environment.
Keywords
control engineering computing; humanoid robots; learning (artificial intelligence); trajectory control; vectors; 51-DOF real humanoid robot; bimanual humanoid motor learning; closed-chain constraint; high-dimensional system; path integral RL framework; reinforcement learning; tool manipulation task; trajectory-model; vector; Humanoid robots; Joints; Predictive models; Standards; Trajectory; Vectors;
fLanguage
English
Publisher
ieee
Conference_Titel
Humanoid Robots (Humanoids), 2013 13th IEEE-RAS International Conference on
Conference_Location
Atlanta, GA
ISSN
2164-0572
Print_ISBN
978-1-4799-2617-6
Type
conf
DOI
10.1109/HUMANOIDS.2013.7030010
Filename
7030010
Link To Document