DocumentCode :
3118269
Title :
Decision theory on dynamic domains nabla derivatives and the Hamilton-Jacobi-Bellman equation
Author :
Seiffertt, John ; Wunsch, Donald C., II ; Sanyal, Suman
Author_Institution :
Dept. of Electr. & Comput. Eng., Missouri Univ. of Sci. & Technol., Rolla, MO
fYear :
2008
fDate :
12-15 Oct. 2008
Firstpage :
3040
Lastpage :
3044
Abstract :
The time scales calculus, which includes the study of the Nabla derivatives, is an emerging key topic due to many multidisciplinary applications. We extend this calculus to approximate dynamic programming. In particular, we investigate application of the Nabla derivative, one of the fundamental dynamic derivatives of time scales. We present a Nabla-derivative based derivation and proof of the Hamilton-Jacobi-Bellman equation, the solution of which is the fundamental problem in the field of dynamic programming. By drawing together the calculus of time scales and the applied area of stochastic control via approximate dynamic programming, we connect two major fields of research.
Keywords :
decision theory; dynamic programming; Hamilton-Jacobi-Bellman equation; Nabla derivative; approximate dynamic programming; decision theory; dynamic domain; stochastic control; time scales calculus; Application software; Calculus; Decision theory; Differential equations; Dynamic programming; Intelligent robots; Learning; Mathematics; Partial differential equations; Stochastic processes; Hamilton-Jacobi-Bellman equation; approximate dynamic programming; reinforcement learning; time scales;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Systems, Man and Cybernetics, 2008. SMC 2008. IEEE International Conference on
Conference_Location :
Singapore
ISSN :
1062-922X
Print_ISBN :
978-1-4244-2383-5
Electronic_ISBN :
1062-922X
Type :
conf
DOI :
10.1109/ICSMC.2008.4811761
Filename :
4811761
Link To Document :
بازگشت