DocumentCode
3572785
Title
A unified approach for semi-Markov decision processes with discounted and average reward criteria
Author
Yanjie Li ; Huijing Wang ; Haoyao Chen
Author_Institution
Shenzhen Grad. Sch., Harbin Inst. of Technol., Shenzhen, China
fYear
2014
Firstpage
1741
Lastpage
1744
Abstract
On the basis of the sensitivity-based optimization, we develop a unified optimization approach for semi-Markov decision processes (SMDPs) with infinite horizon discounted and average reward criteria. We show that the sensitivity formula under average reward criteria is a limitation case of discounted reward criteria. On the basis of the performance sensitivity formulas, we provide a unified formulation for the policy iteration algorithms of semi-Markov decision processes with discounted and average reward criteria.
Keywords
Markov processes; infinite horizon; iterative methods; optimisation; sensitivity; SMDP; average reward criteria; discounted reward criteria; infinite horizon discounted; performance sensitivity formula; policy iteration algorithm; semi-Markov decision process; sensitivity-based optimization; unified optimization approach; Algorithm design and analysis; Educational institutions; Equations; Markov processes; Optimization; Sensitivity; Vectors; SMDPs; performance difference; policy iteration;
fLanguage
English
Publisher
ieee
Conference_Titel
Intelligent Control and Automation (WCICA), 2014 11th World Congress on
Type
conf
DOI
10.1109/WCICA.2014.7052983
Filename
7052983
Link To Document