DocumentCode
2498243
Title
Active exploration by searching for experiments that falsify the computed control policy
Author
Fonteneau, Raphael ; Murphy, Susan A. ; Wehenkel, Louis ; Ernst, Damien
Author_Institution
Dept. of Electr. Eng. & Comput. Sci., Univ. of Liege, Liège, Belgium
fYear
2011
fDate
11-15 April 2011
Firstpage
40
Lastpage
47
Abstract
We propose a strategy for experiment selection - in the context of reinforcement learning - based on the idea that the most interesting experiments to carry out at some stage are those that are the most liable to falsify the current hypothesis about the optimal control policy. We cast this idea in a context where a policy learning algorithm and a model identification method are given a priori. Experiments are selected if, using the learnt environment model, they are predicted to yield a revision of the learnt control policy. Algorithms and simulation results are provided for a deterministic system with discrete action space. They show that the proposed approach is promising.
Keywords
identification; learning (artificial intelligence); optimal control; active exploration; computed control policy; discrete action space; experiment selection; learnt control policy; model identification method; optimal control policy; policy learning algorithm; reinforcement learning; Approximation algorithms; Approximation methods; Heuristic algorithms; Inference algorithms; Optimal control; Prediction algorithms; Predictive models;
fLanguage
English
Publisher
ieee
Conference_Titel
Adaptive Dynamic Programming And Reinforcement Learning (ADPRL), 2011 IEEE Symposium on
Conference_Location
Paris
Print_ISBN
978-1-4244-9887-1
Type
conf
DOI
10.1109/ADPRL.2011.5967364
Filename
5967364
Link To Document