Title :
A framework of an agent planning with reinforcement learning for e-pet
Author :
Sheng-Tzong Cheng ; Tun-Yu Chang ; Chih-Wei Hsu
Author_Institution :
Dept. of Comput. Sci. & Inf. Eng., Nat. Cheng Kung Univ., Tainan, Taiwan
Abstract :
E-pet is an animal-type robot companions, he can be physical or electronic. Reinforcement learning (RL) can be applied to the e-pet. However, the interactive instruction is constituted by complex activities. In this study, we proposed a framework that integrated AI planning technology into RL to generate the solution. In the framework, the e-pet interacts with human and includes two components: environment and agent. The agent exploits AI planning to seek goal state and Markov decision process (MDP) to choose the action and updates each Q-value using Q-learning algorithm. And we proposed the three-level subsumption architecture which including instinct level, perception level, and planning level. We build layers corresponding to each level of competence and can simply add a new layer to an existing set to move to the next higher level of overall competence. We implement the e-pet in a 3D model and train the agent. Experimental result shows that the update of Q-table reduces the number of planning states in the framework.
Keywords :
Markov processes; control engineering computing; human-robot interaction; learning (artificial intelligence); mobile robots; planning (artificial intelligence); 3D model; MDP; Markov decision process; Q-learning algorithm; Q-table; Q-value; RL; agent component; agent training; animal-type robot companions; complex activities; corresponding agent planning level; e-pet; environment component; goal state; instinct level; integrated AI planning technology; interactive instruction; perception level; planning state number reduction; reinforcement learning; three-level subsumption architecture; Learning (artificial intelligence); Markov processes; Planning; Robots; Solid modeling; Three-dimensional displays; AI planning; Markov decision process; e-pet; reinforcement learning; subsumption architecture;
Conference_Titel :
Orange Technologies (ICOT), 2013 International Conference on
Conference_Location :
Tainan
Print_ISBN :
978-1-4673-5934-4
DOI :
10.1109/ICOT.2013.6521220