DocumentCode :
2198644
Title :
On partially observed stochastic shortest path problems
Author :
Patek, Stephen D.
Author_Institution :
Dept. of Syst. & Inf. Eng., Univ. of Virginia, VA, USA
Volume :
5
fYear :
2001
fDate :
2001
Firstpage :
5050
Abstract :
We analyze a class of partially observed stochastic shortest path problems. These are terminating Markov decision process with imperfect state information that evolve on an infinite time horizon and have a total cost criterion. For wellposedness, we make reasonable stochastic shortest path type assumptions: (1) the existence of a policy that guarantees termination with probability one; and (2) the property that any policy that fails to guarantee termination has infinite expected cost from some initial state. We also assume that termination is perfectly recognized. We establish the existence of a stationary optimal policy along with the existence of a unique bounded solution to Bellman´s equation. We also reveal the convergence properties of value and policy iteration. For the case where policies exist that do not guarantee termination, the dynamic programming operator fails to be a contraction mapping with respect to any norm, somewhat complicating the analysis
Keywords :
Markov processes; convergence of numerical methods; dynamic programming; probability; search problems; Markov decision processes; convergence; dynamic programming; optimal search; probability; stochastic shortest path problems; Convergence; Costs; Dynamic programming; Equations; Failure analysis; Information analysis; Shortest path problem; Stochastic processes; Stochastic systems; Systems engineering and theory;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Decision and Control, 2001. Proceedings of the 40th IEEE Conference on
Conference_Location :
Orlando, FL
Print_ISBN :
0-7803-7061-9
Type :
conf
DOI :
10.1109/.2001.981011
Filename :
981011
Link To Document :
بازگشت