Title :
Model-Free reinforcement learning with continuous action in practice
Author :
Degris, T. ; Pilarski, Patrick M. ; Sutton, Richard S.
Author_Institution :
INRIA Bordeaux Sud-Ouest, France
Abstract :
Reinforcement learning methods are often considered as a potential solution to enable a robot to adapt to changes in real time to an unpredictable environment. However, with continuous action, only a few existing algorithms are practical for real-time learning. In such a setting, most effective methods have used a parameterized policy structure, often with a separate parameterized value function. The goal of this paper is to assess such actor-critic methods to form a fully specified practical algorithm. Our specific contributions include 1) developing the extension of existing incremental policy-gradient algorithms to use eligibility traces, 2) an empirical comparison of the resulting algorithms using continuous actions, 3) the evaluation of a gradient-scaling technique that can significantly improve performance. Finally, we apply our actor-critic algorithm to learn on a robotic platform with a fast sensorimotor cycle (10ms). Overall, these results constitute an important step towards practical real-time learning control with continuous action.
Keywords :
gradient methods; learning (artificial intelligence); real-time systems; robots; actor-critic algorithm; continuous action; incremental policy-gradient algorithms; model-free reinforcement learning; parameterized policy structure; real-time learning control; robot; sensorimotor cycle; Learning; Mobile robots; Real-time systems; Robot sensing systems; Standards; Vectors;
Conference_Titel :
American Control Conference (ACC), 2012
Conference_Location :
Montreal, QC
Print_ISBN :
978-1-4577-1095-7
Electronic_ISBN :
0743-1619
DOI :
10.1109/ACC.2012.6315022