DocumentCode :
3507000
Title :
Ex〈α〉: An effective algorithm for continuous actions Reinforcement Learning problems
Author :
Martin H, J.A. ; De Lope, Javier
Author_Institution :
Sist. Informaticos y Computadon, Univ. Complutense de Madrid, Madrid, Spain
fYear :
2009
fDate :
3-5 Nov. 2009
Firstpage :
2063
Lastpage :
2068
Abstract :
In this paper the Ex(α) Reinforcement Learning algorithm is presented. This algorithm is designed to deal with problems where the use of continuous actions have clear advantages over the use of fine grained discrete actions. This new algorithm is derived from a baseline discrete actions algorithm implemented within a kind of κ-nearest neighbors approach in order to produce a probabilistic representation of the input signal to construct robust state descriptions based on a collection (knn) of receptive field units and a probability distribution vector p(knn) over the knn collection. The baseline continuous-space-discrete-actions kNN-TD(λ) algorithm introduces probability traces as the natural adaptation of eligibility traces in the probabilistic context. Later the Ex(α)(κ) algorithm is described as an extension of the baseline algorithms. Finally experimental results are presented for two (not easy) problems such as the Cart-Pole and Helicopter Hovering.
Keywords :
learning (artificial intelligence); probability; baseline algorithms; continuous actions reinforcement learning problems; continuous space discrete actions; discrete actions algorithm; eligibility traces; fine grained discrete actions; k-nearest neighbors; probabilistic representation; probability distribution vector; probability traces; receptive field units; reinforcement learning algorithm;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Industrial Electronics, 2009. IECON '09. 35th Annual Conference of IEEE
Conference_Location :
Porto
ISSN :
1553-572X
Print_ISBN :
978-1-4244-4648-3
Electronic_ISBN :
1553-572X
Type :
conf
DOI :
10.1109/IECON.2009.5415084
Filename :
5415084
Link To Document :
بازگشت