• DocumentCode
    3572785
  • Title

    A unified approach for semi-Markov decision processes with discounted and average reward criteria

  • Author

    Yanjie Li ; Huijing Wang ; Haoyao Chen

  • Author_Institution
    Shenzhen Grad. Sch., Harbin Inst. of Technol., Shenzhen, China
  • fYear
    2014
  • Firstpage
    1741
  • Lastpage
    1744
  • Abstract
    On the basis of the sensitivity-based optimization, we develop a unified optimization approach for semi-Markov decision processes (SMDPs) with infinite horizon discounted and average reward criteria. We show that the sensitivity formula under average reward criteria is a limitation case of discounted reward criteria. On the basis of the performance sensitivity formulas, we provide a unified formulation for the policy iteration algorithms of semi-Markov decision processes with discounted and average reward criteria.
  • Keywords
    Markov processes; infinite horizon; iterative methods; optimisation; sensitivity; SMDP; average reward criteria; discounted reward criteria; infinite horizon discounted; performance sensitivity formula; policy iteration algorithm; semi-Markov decision process; sensitivity-based optimization; unified optimization approach; Algorithm design and analysis; Educational institutions; Equations; Markov processes; Optimization; Sensitivity; Vectors; SMDPs; performance difference; policy iteration;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Intelligent Control and Automation (WCICA), 2014 11th World Congress on
  • Type

    conf

  • DOI
    10.1109/WCICA.2014.7052983
  • Filename
    7052983