• DocumentCode
    776491
  • Title

    Hamilton–Jacobi–Bellman Equations and Approximate Dynamic Programming on Time Scales

  • Author

    Seiffertt, John ; Sanyal, Suman ; Wunsch, Donald C.

  • Author_Institution
    Dept. of Electr. & Comput. Eng., Missouri Univ. of Sci. & Technol., Rolla, MO
  • Volume
    38
  • Issue
    4
  • fYear
    2008
  • Firstpage
    918
  • Lastpage
    923
  • Abstract
    The time scales calculus is a key emerging area of mathematics due to its potential use in a wide variety of multidisciplinary applications. We extend this calculus to approximate dynamic programming (ADP). The core backward induction algorithm of dynamic programming is extended from its traditional discrete case to all isolated time scales. Hamilton-Jacobi-Bellman equations, the solution of which is the fundamental problem in the field of dynamic programming, are motivated and proven on time scales. By drawing together the calculus of time scales and the applied area of stochastic control via ADP, we have connected two major fields of research.
  • Keywords
    calculus; dynamic programming; stochastic systems; Hamilton-Jacobi-Bellman equation; approximate dynamic programming; backward induction algorithm; stochastic control; time scales calculus; Approximate dynamic programming (ADP); Hamilton–Jacobi–Bellman (HJB) equation; Hamilton–Jacobi–Bellman (HJB) equation; dynamic equations; reinforcement learning; time scales; Artificial Intelligence; Computer Simulation; Feedback; Models, Theoretical; Nonlinear Dynamics; Programming, Linear; Systems Theory;
  • fLanguage
    English
  • Journal_Title
    Systems, Man, and Cybernetics, Part B: Cybernetics, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    1083-4419
  • Type

    jour

  • DOI
    10.1109/TSMCB.2008.923532
  • Filename
    4554212