Title :
An analysis of gradient-based policy iteration
Author :
Dankert, James ; Yang, Lei ; Si, Jennie
Author_Institution :
Dept. of Electr. Eng., Arizona State Univ., Tempe, AZ, USA
fDate :
31 July-4 Aug. 2005
Abstract :
Recently, a system theoretic framework for learning and optimization has been developed that shows how many approximate dynamic programming paradigms such as perturbation analysis, Markov decision processes, and reinforcement learning are very closely related. Using this system theoretic framework, a new optimization technique called gradient-based policy iteration (GBPI) has been developed. In this paper, we show how GBPI iteration can be extended to partially observable Markov decision processes (POMDPs). We also develop the value iteration analogue of GBPI and show that this new version of value iteration, extended to POMDPs, not only theoretically acts like value iteration but also does so numerically.
Keywords :
Markov processes; gradient methods; optimisation; system theory; gradient-based policy iteration; partially observable Markov decision process; system theory; value iteration analogue; Algorithm design and analysis; Artificial intelligence; Control systems; Dynamic programming; Electronic mail; Learning; Operations research; Poisson equations; System performance; Terminology;
Conference_Titel :
Neural Networks, 2005. IJCNN '05. Proceedings. 2005 IEEE International Joint Conference on
Print_ISBN :
0-7803-9048-2
DOI :
10.1109/IJCNN.2005.1556399