DocumentCode :
3410074
Title :
Trajectory-model-based reinforcement learning: Application to bimanual humanoid motor learning with a closed-chain constraint
Author :
Sugimoto, Norikazu ; Morimoto, Jun
Author_Institution :
Dept. of Brain Machine Interface, Nat. Inst. of Inf. & Commun. Technol., Suita, Japan
fYear :
2013
fDate :
15-17 Oct. 2013
Firstpage :
429
Lastpage :
434
Abstract :
We propose a reinforcement learning (RL) framework to improve policies for a high-dimensional system through fewer interactions with real environments than standard RL methods. In our learning framework, we first use off-line simulations to improve the controller parameters with an approximated environment model to generate samples along locally optimized trajectories. We then use the approximated dynamics to improve the performance of a tool manipulation task in a path integral RL framework, which updates a policy from the sampled trajectories of the state and action vectors and the cost. In this study, we apply our proposed method to a bimanual humanoid motor learning task in which we need to explicitly consider a closed-chain constraint. We show that a 51-DOF real humanoid robot can learn to manipulate a rod to hit via-points using both arms within 36 interactions in a real environment.
Keywords :
control engineering computing; humanoid robots; learning (artificial intelligence); trajectory control; vectors; 51-DOF real humanoid robot; bimanual humanoid motor learning; closed-chain constraint; high-dimensional system; path integral RL framework; reinforcement learning; tool manipulation task; trajectory-model; vector; Humanoid robots; Joints; Predictive models; Standards; Trajectory; Vectors;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Humanoid Robots (Humanoids), 2013 13th IEEE-RAS International Conference on
Conference_Location :
Atlanta, GA
ISSN :
2164-0572
Print_ISBN :
978-1-4799-2617-6
Type :
conf
DOI :
10.1109/HUMANOIDS.2013.7030010
Filename :
7030010
Link To Document :
بازگشت