DocumentCode :
3312673
Title :
Using dimensionality reduction to exploit constraints in reinforcement learning
Author :
Bitzer, Sebastian ; Howard, Matthew ; Vijayakumar, Sethu
Author_Institution :
Sch. of Inf., Univ. of Edinburgh, Edinburgh, UK
fYear :
2010
fDate :
18-22 Oct. 2010
Firstpage :
3219
Lastpage :
3225
Abstract :
Reinforcement learning in the high-dimensional, continuous spaces typical in robotics, remains a challenging problem. To overcome this challenge, a popular approach has been to use demonstrations to find an appropriate initialisation of the policy in an attempt to reduce the number of iterations needed to find a solution. Here, we present an alternative way to incorporate prior knowledge from demonstrations of individual postures into learning, by extracting the inherent problem structure to find an efficient state representation. In particular, we use probabilistic, nonlinear dimensionality reduction to capture latent constraints present in the data. By learning policies in the learnt latent space, we are able to solve the planning problem in a reduced space that automatically satisfies task constraints. As shown in our experiments, this reduces the exploration needed and greatly accelerates the learning. We demonstrate our approach for learning a bimanual reaching task on the 19-DOF KHR-1HV humanoid.
Keywords :
humanoid robots; learning (artificial intelligence); path planning; robot kinematics; 19-DOF KHR-1HV humanoid; continuous spaces; dimensionality reduction; latent constraints; learning policies; nonlinear dimensionality reduction; planning problem; probabilistic reduction; reinforcement learning; robotics; state representation;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Intelligent Robots and Systems (IROS), 2010 IEEE/RSJ International Conference on
Conference_Location :
Taipei
ISSN :
2153-0858
Print_ISBN :
978-1-4244-6674-0
Type :
conf
DOI :
10.1109/IROS.2010.5650243
Filename :
5650243
Link To Document :
بازگشت