Title :
A study on use of prior information for acceleration of reinforcement learning
Author :
Terashima, Kento ; Murata, Junichi
Author_Institution :
Dept. of Electr. & Electron. Eng., Kyushu Univ., Fukuoka, Japan
Abstract :
Reinforcement learning is a method with which an agent learns appropriate response for solving problems by trial-and-error. The advantage is that reinforcement learning can be applied to unknown or uncertain problems. But instead, there is a drawback that this method needs a long time to solve the problem because of trial-and-error. If there is prior information about the environment, some of trial-and-error can be spared and the learning can take a shorter time. The prior information provided by a human designer can be wrong because of uncertainties in the problems. If the wrong prior information is used, there can be bad effects such as failure to get the optimal policy and slowing down of reinforcement learning. We propose to control use of the prior information to suppress the bad effects. The agent forgets the prior information gradually by multiplying a forgetting factor while it learns the better policy. We apply the proposed method to a couple of testbed environments and a number of types of prior information. The method shows the good results in terms of both the learning speed and the quality of obtained policies.
Keywords :
learning (artificial intelligence); multi-agent systems; problem solving; uncertainty handling; agent learning; optimal policy; prior information; problem solving; reinforcement learning; trial-and-error; uncertainties; Acceleration; Educational institutions; Focusing; Humans; Learning; Learning systems; Trajectory; exploring visit; forgetting factor; option; prior information; reinforcement learning;
Conference_Titel :
SICE Annual Conference (SICE), 2011 Proceedings of
Conference_Location :
Tokyo
Print_ISBN :
978-1-4577-0714-8