DocumentCode :
607751
Title :
Sequential Monte Carlo samplers for model-based reinforcement learning
Author :
Sonmez, O. ; Cemgil, A.T.
Author_Institution :
Bilgisayar Muhendisligi Bolumu, Bogazici Univ., Istanbul, Turkey
fYear :
2013
fDate :
24-26 April 2013
Firstpage :
1
Lastpage :
4
Abstract :
Reinforcement learning problems are generally solved by using fixed-point iterations that converge to the suboptimal solutions of Bellman equations. However, it is also possible to formalize this problem as an equivalent likelihood maximization problem and employ probabilistic inference methods. We proposed an expectation-maximization algorithm that utilizes sequential Monte Carlo samplers with Metropolis-Hastings kernels in its expectation step to solve the model-based version. Then, we evaluate our algorithm on mountain-car problem which is a benchmark reinforcement learning problem.
Keywords :
Monte Carlo methods; expectation-maximisation algorithm; learning (artificial intelligence); Bellman equations; equivalent likelihood maximization problem; expectation-maximization algorithm; fixed-point iterations; model-based reinforcement learning; probabilistic inference methods; sequential Monte Carlo samplers; Electronic mail; Learning (artificial intelligence); Markov processes; Mathematical model; Monte Carlo methods; Presses; Probabilistic logic; Expectation-Maximization; Markov Decision Processes; Metropolis-Hastings; Reinforcement Learning; Sequential Monte Carlo Samplers;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Signal Processing and Communications Applications Conference (SIU), 2013 21st
Conference_Location :
Haspolat
Print_ISBN :
978-1-4673-5562-9
Electronic_ISBN :
978-1-4673-5561-2
Type :
conf
DOI :
10.1109/SIU.2013.6531412
Filename :
6531412
Link To Document :
بازگشت