Title :
Trading off perception with internal state: reinforcement learning and analysis of Q-Elman networks in a Markovian task
Author :
Bakker, Bram ; van der Voort van der Kleif, G.
Author_Institution :
Unit of Exp. & Theor. Psychol., Leiden Univ., Netherlands
Abstract :
A Markovian reinforcement learning task can be dealt with by learning a direct mapping from states to actions or values, or from state-action pairs to values. However, this may involve a difficult pattern recognition problem when the state space is large. This paper shows that using internal state, called “supportive state”, may alleviate this problem presenting an argument against the tendency to almost automatically use a direct mapping when the task is Markovian. This point is demonstrated in simulation experiments of an agent controlled by a neural network capable of learning the strategy of direct mapping as well as internal state, combining Q(λ) learning and recurrent neural networks in a new way. The trade-off between the two strategies is investigated in more detail, focusing particularly on border cases
Keywords :
Markov processes; learning (artificial intelligence); pattern recognition; recurrent neural nets; state-space methods; Markov processes; Q-Elman networks; direct mapping; internal state; pattern recognition; recurrent neural networks; reinforcement learning; state space; Automatic control; Feedforward neural networks; History; Intelligent agent; Learning; Neural networks; Pattern recognition; Psychology; Recurrent neural networks; State-space methods;
Conference_Titel :
Neural Networks, 2000. IJCNN 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on
Conference_Location :
Como
Print_ISBN :
0-7695-0619-4
DOI :
10.1109/IJCNN.2000.861306