DocumentCode :
3432655
Title :
Nonlinear two-player zero-sum game approximate solution using a Policy Iteration algorithm
Author :
Johnson, M. ; Bhasin, S. ; Dixon, W.E.
Author_Institution :
Dept. of Mechanical and Aerospace Engineering, University of Florida, Gainesville, 32611, USA
fYear :
2011
fDate :
12-15 Dec. 2011
Firstpage :
142
Lastpage :
147
Abstract :
An approximate online solution is developed for a two-player zero-sum game subject to continuous-time nonlinear uncertain dynamics and an infinite horizon quadratic cost. A novel actor-critic-identifier (ACI) structure is used to implement the Policy Iteration (PI) algorithm, wherein a robust dynamic neural network (DNN) is used to asymptotically identify the uncertain system, and a critic NN is used to approximate the value function. The weight update laws for the critic NN are generated using a gradient-descent method based on a modified temporal difference error, which is independent of the system dynamics. This method finds approximations of the optimal value function, and the saddle point feedback control policies. These policies are computed using the critic NN and the identifier DNN and guarantee uniformly ultimately bounded (UUB) stability of the closed-loop system. The actor, critic and identifier structures are implemented in real-time, continuously and simultaneously.
Keywords :
Approximation algorithms; Approximation methods; Artificial neural networks; Equations; Game theory; Games; Heuristic algorithms;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Decision and Control and European Control Conference (CDC-ECC), 2011 50th IEEE Conference on
Conference_Location :
Orlando, FL, USA
ISSN :
0743-1546
Print_ISBN :
978-1-61284-800-6
Electronic_ISBN :
0743-1546
Type :
conf
DOI :
10.1109/CDC.2011.6160778
Filename :
6160778
Link To Document :
بازگشت