Title :
Q-learning policies for a single agent foraging tasks
Author :
Mohan, Yogeswaran ; Ponnambalam, S.G.
Author_Institution :
Sch. of Eng., Monash Univ., Petaling Jaya, Malaysia
Abstract :
Policies play an important role in balancing the trade-off between exploration and exploitation problem in q-learning. Pure exploration degrades the performance of the q-learning but increases the flexibility to adapt in a dynamic environment. On the other hand pure exploitation drives the learning process to locally optimal solutions. In this paper, a single agent foraging task has been modeled incorporating the available policies reported in the open literature to address the exploration and exploitation issues. Policies namely greedy, e-greedy, Boltzmann distribution, Simulated An-nealing(SA)algorithm and random search are used to study their performances in the foraging task and the results are presented.
Keywords :
greedy algorithms; learning (artificial intelligence); simulated annealing; Boltzmann distribution; Q-learning policies; e-greedy; optimal solutions; simulated annealing; single agent foraging tasks; Equations; Global Positioning System; Hardware; Marine technology; Navigation; Radio transmitters; Radiofrequency identification; Robots; Simultaneous localization and mapping; Wireless LAN;
Conference_Titel :
Mechatronics and its Applications (ISMA), 2010 7th International Symposium on
Conference_Location :
Sharjah
Print_ISBN :
978-1-4244-6665-8