Title :
A hybrid architecture for hierarchical reinforcement learning
Author_Institution :
Dept. of Comput. Sci. Eng., Texas Univ., Arlington, TX, USA
Abstract :
Autonomous robot systems operating in the real world have to be able to learn new tasks and environmental conditions without the need for an outside teacher. While reinforcement learning represents a good formalism to achieve this, its long learning times and need for extensive exploration often make it impracticable for online learning on complex systems. The hybrid architecture presented in this paper addresses this issue by applying reinforcement learning on top of an automatically derived abstract discrete event dynamic system (DEDS) supervisor. This reduces the problem of policy acquisition within this approach to learning to coordinate a set of closed-loop control strategies in order to perform a given task. Besides dramatically reducing the complexity of the learning task this framework also permits the incorporation of a priori knowledge and facilitates the inclusion of learned policies as actions in order to transfer skills to new task domains. To demonstrate the applicability of this approach, the architecture is used to learn locomotion gaits on a four-legged robot platform
Keywords :
closed loop systems; computational complexity; discrete event systems; learning (artificial intelligence); mobile robots; DEDS supervisor; autonomous robot systems; closed-loop control strategies; complex systems; discrete event dynamic system supervisor; environmental condition learning; four-legged robot platform; hierarchical reinforcement learning; hybrid architecture; locomotion gaits; online learning; policy acquisition; Automatic control; Bridges; Computer architecture; Computer science; Control systems; Educational robots; Learning systems; Optimal control; Robot kinematics; State-space methods;
Conference_Titel :
Robotics and Automation, 2000. Proceedings. ICRA '00. IEEE International Conference on
Conference_Location :
San Francisco, CA
Print_ISBN :
0-7803-5886-4
DOI :
10.1109/ROBOT.2000.845170