DocumentCode :
954976
Title :
A simultaneous perturbation stochastic approximation-based actor-critic algorithm for Markov decision processes
Author :
Bhatnagar, Shalabh ; Kumar, Shishir
Author_Institution :
Dept. of Comput. Sci. & Autom., Indian Inst. of Sci., Bangalore, India
Volume :
49
Issue :
4
fYear :
2004
fDate :
4/1/2004 12:00:00 AM
Firstpage :
592
Lastpage :
598
Abstract :
A two-timescale simulation-based actor-critic algorithm for solution of infinite horizon Markov decision processes with finite state and compact action spaces under the discounted cost criterion is proposed. The algorithm does gradient search on the slower timescale in the space of deterministic policies and uses simultaneous perturbation stochastic approximation-based estimates. On the faster scale, the value function corresponding to a given stationary policy is updated and averaged over a fixed number of epochs (for enhanced performance). The proof of convergence to a locally optimal policy is presented. Finally, numerical experiments using the proposed algorithm on flow control in a bottleneck link using a continuous time queueing model are shown.
Keywords :
Markov processes; approximation theory; continuous time systems; convergence; gradient methods; perturbation techniques; queueing theory; search problems; Markov decision processes; actor-critic algorithm; compact action spaces; continuous time queueing model; convergence; discounted cost criterion; finite state; flow control; gradient search; simultaneous perturbation stochastic approximation; Automatic control; Circuits; Control system synthesis; Control systems; Feedback; Modules (abstract algebra); Multidimensional systems; Polynomials; Stochastic processes; System analysis and design;
fLanguage :
English
Journal_Title :
Automatic Control, IEEE Transactions on
Publisher :
ieee
ISSN :
0018-9286
Type :
jour
DOI :
10.1109/TAC.2004.825622
Filename :
1284724
Link To Document :
بازگشت