DocumentCode
395547
Title
Anticipative reinforcement learning
Author
Maire, Frederic
Author_Institution
Sch. of Comput. Sci. & Software Eng., Queensland Univ. of Technol., Brisbane, Qld., Australia
Volume
3
fYear
2002
fDate
18-22 Nov. 2002
Firstpage
1428
Abstract
This paper introduces anticipative reinforcement learning (ARL), a method that addresses the problem of the breakdown of value based algorithms for problems with small time steps and continuous action and state spaces when the algorithms are implemented with neural networks. In ARL, an agent is made of three components; the actor, the critic and the model (the model is as in Dyna but we use it differently). The main originality of ARL lies in the action selection process; the agent builds a set of candidate actions that includes the action recommended by the actor plus some random actions. Once the set of candidate actions is built, the candidate actions are ranked by considering what would happen if these actions were taken and followed by a sequence of actions using only the current policy (anticipation using iteratively the model with a finite look-ahead). We demonstrate the benefits of looking ahead with experiments on a Khepera robot.
Keywords
function approximation; generalisation (artificial intelligence); learning (artificial intelligence); mobile robots; neural nets; state-space methods; Khepera robot; anticipative reinforcement learning; function approximation; generalisation; neural networks; state spaces; Australia; Books; Electric breakdown; Laboratories; Learning; Robots; Software algorithms; Software engineering; Space technology; State-space methods;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Information Processing, 2002. ICONIP '02. Proceedings of the 9th International Conference on
Print_ISBN
981-04-7524-1
Type
conf
DOI
10.1109/ICONIP.2002.1202856
Filename
1202856
Link To Document