DocumentCode :
2162113
Title :
Importance sampling for model-based reinforcement learning
Author :
Sönmez, Orhan ; Cemgil, A. Taylan
Author_Institution :
Bilgisayar Muhendisligi Bolumu, Bogazici Univ., Istanbul, Turkey
fYear :
2012
fDate :
18-20 April 2012
Firstpage :
1
Lastpage :
4
Abstract :
Most of the state-of-the-art reinforcement learning algorithms are based on Bellman equations and make use of fixed-point iteration methods to converge to suboptimal solutions. However, some of the recent approaches transform the reinforcement learning problem into an equivalent likelihood maximization problem with using appropriate graphical models. Hence, it allows the adoption of probabilistic inference methods. Here, we propose an expectation-maximization method that employs importance sampling in its E-step in order to estimate the likelihood and then to determine the optimal policy.
Keywords :
expectation-maximisation algorithm; importance sampling; inference mechanisms; learning (artificial intelligence); Bellman equation; E-step; equivalent likelihood maximization problem; expectation-maximization method; fixed-point iteration method; graphical models; importance sampling; model-based reinforcement learning; optimal policy; probabilistic inference method; Abstracts; Inference algorithms; Learning; Machine learning; Markov processes; Monte Carlo methods; Probabilistic logic;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Signal Processing and Communications Applications Conference (SIU), 2012 20th
Conference_Location :
Mugla
Print_ISBN :
978-1-4673-0055-1
Electronic_ISBN :
978-1-4673-0054-4
Type :
conf
DOI :
10.1109/SIU.2012.6204703
Filename :
6204703
Link To Document :
بازگشت