Title :
Model-based reinforcement learning for on-line feedback-Nash equilibrium solution of N-player nonzero-sum differential games
Author :
Kamalapurkar, Rushikesh ; Klotz, J. ; Dixon, Warren E.
Author_Institution :
Dept. of Mech. & Aerosp. Eng., Univ. of Florida, Gainesville, FL, USA
Abstract :
This paper presents a concurrent learning-based actor-critic-identifier architecture to obtain an approximate feedback-Nash equilibrium solution to a deterministic, continuous-time, and infinite-horizon N-player nonzero-sum differential game on-line, without requiring persistence of excitation (PE), for non-linear control-affine systems. Convergence of the developed control policies to neighborhoods of the feedback-Nash equilibrium policies is established under a sufficient rank condition. Simulation results are presented to demonstrate the performance of the developed technique.
Keywords :
affine transforms; continuous time systems; differential games; infinite horizon; learning (artificial intelligence); nonlinear control systems; N-player nonzero-sum differential games; approximate feedback-Nash equilibrium solution; concurrent learning-based actor-critic-identifier architecture; continuous-time differential game; deterministic differential game; developed control policy; feedback-Nash equilibrium policy; infinite-horizon N-player nonzero-sum differential game online; model-based reinforcement learning; nonlinear control-affine system; online feedback-Nash equilibrium solution; persistence of excitation; sufficient rank condition; Artificial neural networks; Convergence; Function approximation; Games; Least squares approximations; Trajectory; Adaptive systems; Learning; Nonlinear systems;
Conference_Titel :
American Control Conference (ACC), 2014
Conference_Location :
Portland, OR
Print_ISBN :
978-1-4799-3272-6
DOI :
10.1109/ACC.2014.6859092