DocumentCode :
391016
Title :
Convergence of numerical optimal feedback policies for deterministic optimal control problems
Author :
Dupuis, Paul ; Szpiro, Adam
Author_Institution :
Div. of Appl. Math., Brown Univ., Providence, RI, USA
Volume :
3
fYear :
2002
fDate :
10-13 Dec. 2002
Firstpage :
3138
Abstract :
We consider a Markov chain based numerical approximation method for a class of deterministic nonlinear optimal control problems. It is known that methods of this type yield convergent approximations to the value function on the entire domain. These results do not easily extend to the optimal control, which need not be uniquely defined on the entire domain. There are, however, regions of strong regularity on which the optimal control is well defined and smooth. Typically, the union of these regions is open and dense in the domain. Using probabilistic methods, we prove that on the regions of strong regularity, the Markov chain method yields a convergent sequence of approximations to the optimal feedback control. The result is illustrated with several examples.
Keywords :
Markov processes; approximation theory; convergence of numerical methods; feedback; nonlinear control systems; optimal control; probability; Markov chain based numerical approximation method; convergence; convergent approximations; deterministic optimal control problems; nonlinear control problems; numerical optimal feedback policies; probabilistic methods; strong regularity; Approximation methods; Convergence of numerical methods; Cost function; Feedback control; Finite difference methods; Mathematics; Optimal control; Stochastic processes; Symmetric matrices; US Government;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Decision and Control, 2002, Proceedings of the 41st IEEE Conference on
ISSN :
0191-2216
Print_ISBN :
0-7803-7516-5
Type :
conf
DOI :
10.1109/CDC.2002.1184352
Filename :
1184352
Link To Document :
بازگشت