DocumentCode :
2276906
Title :
Reinforcement Learning in Humanoid Robotics Dusko Katic
fYear :
2006
fDate :
25-27 Sept. 2006
Firstpage :
85
Lastpage :
85
Abstract :
Summary form only given. Dynamic bipedal walking is difficult to learn because combinatorial explosion in order to optimize performance in every possible configuration of the robot, uncertainties of the robot dynamics that must be only experimentally validated, and because coping with dynamic discontinuities caused by collisions with the ground and with the problem of delayed reward-torques applied at one time may have an effect on the performance many steps into the future. The detailed and precise training data for learning is often hard to obtain or may not be available in the process of biped control synthesis. Since no exact teaching information is available, this is a typical reinforcement learning problem and the failure signal serves as the reinforcement signal. Reinforcement learning (RL) offers one of the most general framework to humanoid robotics towards true autonomy and versatility. Various straightforward and hybrid intelligent control algorithms based RL for active and passive biped locomotion is presented. The proposed reinforcement learning algorithms is based on two different learning structures: actor-critic architecture and Q-learning structures. Also, RL algorithms can use numerical and fuzzy evaluative feedback information for external reinforcement. The proposed RL algorithms use the learning elements that consist of various types of neural networks, fuzzy logic nets or fuzzy-neuro networks with focus on fast convergence properties and small number of learning trials
Keywords :
fuzzy neural nets; humanoid robots; intelligent control; learning (artificial intelligence); legged locomotion; robot dynamics; Q-learning structures; active biped locomotion; actor-critic architecture; dynamic bipedal walking; fuzzy logic nets; fuzzy-neuro networks; humanoid robotics; hybrid intelligent control algorithms based RL; neural networks; passive biped locomotion; reinforcement learning algorithms; robot dynamics uncertainties; Delay effects; Education; Explosions; Humanoid robots; Learning; Legged locomotion; Process control; Signal synthesis; Training data; Uncertainty;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Network Applications in Electrical Engineering, 2006. NEUREL 2006. 8th Seminar on
Conference_Location :
Belgrade, Serbia & Montenegro
Print_ISBN :
1-4244-0433-9
Electronic_ISBN :
1-4244-0433-9
Type :
conf
DOI :
10.1109/NEUREL.2006.341182
Filename :
4147170
Link To Document :
بازگشت