• DocumentCode
    2162113
  • Title

    Importance sampling for model-based reinforcement learning

  • Author

    Sönmez, Orhan ; Cemgil, A. Taylan

  • Author_Institution
    Bilgisayar Muhendisligi Bolumu, Bogazici Univ., Istanbul, Turkey
  • fYear
    2012
  • fDate
    18-20 April 2012
  • Firstpage
    1
  • Lastpage
    4
  • Abstract
    Most of the state-of-the-art reinforcement learning algorithms are based on Bellman equations and make use of fixed-point iteration methods to converge to suboptimal solutions. However, some of the recent approaches transform the reinforcement learning problem into an equivalent likelihood maximization problem with using appropriate graphical models. Hence, it allows the adoption of probabilistic inference methods. Here, we propose an expectation-maximization method that employs importance sampling in its E-step in order to estimate the likelihood and then to determine the optimal policy.
  • Keywords
    expectation-maximisation algorithm; importance sampling; inference mechanisms; learning (artificial intelligence); Bellman equation; E-step; equivalent likelihood maximization problem; expectation-maximization method; fixed-point iteration method; graphical models; importance sampling; model-based reinforcement learning; optimal policy; probabilistic inference method; Abstracts; Inference algorithms; Learning; Machine learning; Markov processes; Monte Carlo methods; Probabilistic logic;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Signal Processing and Communications Applications Conference (SIU), 2012 20th
  • Conference_Location
    Mugla
  • Print_ISBN
    978-1-4673-0055-1
  • Electronic_ISBN
    978-1-4673-0054-4
  • Type

    conf

  • DOI
    10.1109/SIU.2012.6204703
  • Filename
    6204703