DocumentCode :
2179196
Title :
On step sizes, stochastic shortest paths, and survival probabilities in Reinforcement Learning
Author :
Gosavi, Abhijit
Author_Institution :
Dept. of Eng. Manage. & Syst. Eng., Missouri Univ. of Sci. & Technol., Rolla, MO, USA
fYear :
2008
fDate :
7-10 Dec. 2008
Firstpage :
525
Lastpage :
531
Abstract :
Reinforcement learning (RL) is a simulation-based technique useful in solving Markov decision processes if their transition probabilities are not easily obtainable or if the problems have a very large number of states. We present an empirical study of (i) the effect of step-sizes (learning rules) in the convergence of RL algorithms, (ii) stochastic shortest paths in solving average reward problems via RL, and (iii) the notion of survival probabilities (downside risk) in RL. We also study the impact of step sizes when function approximation is combined with RL. Our experiments yield some interesting insights that will be useful in practice when RL algorithms are implemented within simulators.
Keywords :
Markov processes; function approximation; learning (artificial intelligence); Markov decision processes; function approximation; reinforcement learning; stochastic shortest paths; survival probabilities; Approximation algorithms; Convergence; Function approximation; Learning; Modeling; Polynomials; Research and development management; Stochastic processes; Stochastic systems; Systems engineering and theory;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Simulation Conference, 2008. WSC 2008. Winter
Conference_Location :
Austin, TX
Print_ISBN :
978-1-4244-2707-9
Electronic_ISBN :
978-1-4244-2708-6
Type :
conf
DOI :
10.1109/WSC.2008.4736109
Filename :
4736109
Link To Document :
بازگشت