DocumentCode :
353244
Title :
Trading off perception with internal state: reinforcement learning and analysis of Q-Elman networks in a Markovian task
Author :
Bakker, Bram ; van der Voort van der Kleif, G.
Author_Institution :
Unit of Exp. & Theor. Psychol., Leiden Univ., Netherlands
Volume :
3
fYear :
2000
fDate :
2000
Firstpage :
213
Abstract :
A Markovian reinforcement learning task can be dealt with by learning a direct mapping from states to actions or values, or from state-action pairs to values. However, this may involve a difficult pattern recognition problem when the state space is large. This paper shows that using internal state, called “supportive state”, may alleviate this problem presenting an argument against the tendency to almost automatically use a direct mapping when the task is Markovian. This point is demonstrated in simulation experiments of an agent controlled by a neural network capable of learning the strategy of direct mapping as well as internal state, combining Q(λ) learning and recurrent neural networks in a new way. The trade-off between the two strategies is investigated in more detail, focusing particularly on border cases
Keywords :
Markov processes; learning (artificial intelligence); pattern recognition; recurrent neural nets; state-space methods; Markov processes; Q-Elman networks; direct mapping; internal state; pattern recognition; recurrent neural networks; reinforcement learning; state space; Automatic control; Feedforward neural networks; History; Intelligent agent; Learning; Neural networks; Pattern recognition; Psychology; Recurrent neural networks; State-space methods;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2000. IJCNN 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on
Conference_Location :
Como
ISSN :
1098-7576
Print_ISBN :
0-7695-0619-4
Type :
conf
DOI :
10.1109/IJCNN.2000.861306
Filename :
861306
Link To Document :
بازگشت