DocumentCode :
171845
Title :
Curiosity-driven exploration in reinforcement learning
Author :
Gregor, Matthias ; Spalek, Juraj
Author_Institution :
Dept. of Control & Inf. Syst., Univ. of Zilina, Zilina, Slovakia
fYear :
2014
fDate :
19-20 May 2014
Firstpage :
435
Lastpage :
440
Abstract :
The paper elaborates upon a prior proposal for a novelty detector based on an artificial neural network forecaster. In the former paper, the novelty-based motivational signal was used in place of more conventional techniques (such as the ε-greedy policy, or the softmax policy) to drive exploration, in the context of V-learning. The current paper provides a more comprehensive study of such handling of the exploration vs. exploitation trade-off. It also studies the various problems concerning application of the approach to SARSA, and Q-learning. Also, and with the same goal in mind, the paper presents several advances upon the original design.
Keywords :
learning (artificial intelligence); neural nets; Q-learning; SARSA; V-learning; artificial neural network forecaster; curiosity-driven exploration; novelty detector; novelty-based motivational signal; reinforcement learning; Switches;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
ELEKTRO, 2014
Conference_Location :
Rajecke Teplice
Print_ISBN :
978-1-4799-3720-2
Type :
conf
DOI :
10.1109/ELEKTRO.2014.6848933
Filename :
6848933
Link To Document :
بازگشت