• DocumentCode
    337016
  • Title

    Minimax-based reinforcement learning with state aggregation

  • Author

    Jiang, Guofei ; Wu, Cang-Pu ; Cybenko, George

  • Author_Institution
    Dept. of Autom. Control, Beijing Inst. of Technol., China
  • Volume
    2
  • fYear
    1998
  • fDate
    16-18 Dec 1998
  • Firstpage
    1236
  • Abstract
    One of the most important issues in scaling up reinforcement learning for practical problems is how to represent and store cost-to-go functions with more compact representations than lookup tables. We address the issue of combining the simple function approximation method-state aggregation with minimax-based reinforcement learning algorithms and present the convergence theory for online Q-hat-learning with state aggregation. Some empirical results are also included
  • Keywords
    Markov processes; aggregation; convergence; decision theory; function approximation; learning (artificial intelligence); convergence theory; cost-to-go functions; function approximation method; minimax-based reinforcement learning; online Q-hat-learning; state aggregation; Approximation algorithms; Convergence; Costs; Dynamic programming; Educational institutions; Function approximation; Learning; Minimax techniques; Stochastic processes; Table lookup;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Decision and Control, 1998. Proceedings of the 37th IEEE Conference on
  • Conference_Location
    Tampa, FL
  • ISSN
    0191-2216
  • Print_ISBN
    0-7803-4394-8
  • Type

    conf

  • DOI
    10.1109/CDC.1998.758445
  • Filename
    758445