DocumentCode :
2717652
Title :
Reinforcement Learning in Continuous Action Spaces
Author :
Van Hasselt, Hado ; Wiering, Marco A.
Author_Institution :
Dept. of Inf. & Comput. Sci., Utrecht Univ.
fYear :
2007
fDate :
1-5 April 2007
Firstpage :
272
Lastpage :
279
Abstract :
Quite some research has been done on reinforcement learning in continuous environments, but the research on problems where the actions can also be chosen from a continuous space is much more limited. We present a new class of algorithms named continuous actor critic learning automaton (CACLA) that can handle continuous states and actions. The resulting algorithm is straightforward to implement. An experimental comparison is made between this algorithm and other algorithms that can handle continuous action spaces. These experiments show that CACLA performs much better than the other algorithms, especially when it is combined with a Gaussian exploration method
Keywords :
continuous systems; learning (artificial intelligence); learning automata; continuous action space; continuous actor critic learning automaton; continuous state; reinforcement learning; Books; Computational modeling; Dynamic programming; Intelligent systems; Learning automata; Physics computing; Telephony;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Approximate Dynamic Programming and Reinforcement Learning, 2007. ADPRL 2007. IEEE International Symposium on
Conference_Location :
Honolulu, HI
Print_ISBN :
1-4244-0706-0
Type :
conf
DOI :
10.1109/ADPRL.2007.368199
Filename :
4220844
Link To Document :
بازگشت