DocumentCode :
2498014
Title :
Parametric value function approximation: A unified view
Author :
Geist, Matthieu ; Pietquin, Olivier
Author_Institution :
IMS Res. Group, Supelec, Metz, France
fYear :
2011
fDate :
11-15 April 2011
Firstpage :
9
Lastpage :
16
Abstract :
Reinforcement learning (RL) is a machine learning answer to the optimal control problem. It consists of learning an optimal control policy through interactions with the system to be controlled, the quality of this policy being quantified by the so-called value function. An important RL subtopic is to approximate this function when the system is too large for an exact representation. This survey reviews and unifies state of the art methods for parametric value function approximation by grouping them into three main categories: bootstrapping, residuals and projected fixed-point approaches. Related algorithms are derived by considering one of the associated cost functions and a specific way to minimize it, almost always a stochastic gradient descent or a recursive least-squares approach.
Keywords :
function approximation; gradient methods; learning (artificial intelligence); least squares approximations; bootstrapping approach; machine learning; optimal control policy; parametric value function approximation; projected fixed-point approach; recursive least-squares approach; reinforcement learning; residuals approach; stochastic gradient descent approach; Approximation algorithms; Cost function; Equations; Function approximation; Mathematical model; Prediction algorithms; Stochastic processes; Reinforcement learning; survey; value function approximation;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Adaptive Dynamic Programming And Reinforcement Learning (ADPRL), 2011 IEEE Symposium on
Conference_Location :
Paris
Print_ISBN :
978-1-4244-9887-1
Type :
conf
DOI :
10.1109/ADPRL.2011.5967355
Filename :
5967355
Link To Document :
بازگشت