Title :
Solving Bellman´s equation by solving method of continuity
Author_Institution :
Dept. of Comput. Sci., Massachusetts Univ., Amherst, MA, USA
fDate :
29 June-1 July 1994
Abstract :
It is known that policy iteration can be identified with Newton´s method (and value iteration with successive approximation) for solving Bellman´s optimality equation, which for Markov decision problems takes the form: for i=1, N F(v(i))=maxu∈U/[r~u(i)+αΣj=1NPu(i,j)v(j)]-v(i)=0. One is naturally led to consider what new computational methods might be suggested by adopting alternative root-finding procedures. This paper summarises an investigation of the consequences of adopting one particular root-finding scheme called the method of continuity (also known as the method of imbedding or homotopy).
Keywords :
Markov processes; Newton method; decision theory; optimisation; Bellman´s optimality equation; Markov decision problems; Newton´s method; continuity method; homotopy; imbedding; policy iteration; root-finding procedures; successive approximation; value iteration; Computer science; Nonlinear equations;
Conference_Titel :
American Control Conference, 1994
Print_ISBN :
0-7803-1783-1
DOI :
10.1109/ACC.1994.735045