Title :
Sequential Monte Carlo samplers for model-based reinforcement learning
Author :
Sonmez, O. ; Cemgil, A.T.
Author_Institution :
Bilgisayar Muhendisligi Bolumu, Bogazici Univ., Istanbul, Turkey
Abstract :
Reinforcement learning problems are generally solved by using fixed-point iterations that converge to the suboptimal solutions of Bellman equations. However, it is also possible to formalize this problem as an equivalent likelihood maximization problem and employ probabilistic inference methods. We proposed an expectation-maximization algorithm that utilizes sequential Monte Carlo samplers with Metropolis-Hastings kernels in its expectation step to solve the model-based version. Then, we evaluate our algorithm on mountain-car problem which is a benchmark reinforcement learning problem.
Keywords :
Monte Carlo methods; expectation-maximisation algorithm; learning (artificial intelligence); Bellman equations; equivalent likelihood maximization problem; expectation-maximization algorithm; fixed-point iterations; model-based reinforcement learning; probabilistic inference methods; sequential Monte Carlo samplers; Electronic mail; Learning (artificial intelligence); Markov processes; Mathematical model; Monte Carlo methods; Presses; Probabilistic logic; Expectation-Maximization; Markov Decision Processes; Metropolis-Hastings; Reinforcement Learning; Sequential Monte Carlo Samplers;
Conference_Titel :
Signal Processing and Communications Applications Conference (SIU), 2013 21st
Conference_Location :
Haspolat
Print_ISBN :
978-1-4673-5562-9
Electronic_ISBN :
978-1-4673-5561-2
DOI :
10.1109/SIU.2013.6531412