DocumentCode :
2065158
Title :
Situated perception of a partner robot based on neuro-fuzzy computing
Author :
Kubota, Naoyuki ; Nishida, Kenichiro
Author_Institution :
Dept. of Syst. Design, Tokyo Metropolitan Univ., Japan
fYear :
2005
fDate :
12-15 June 2005
Firstpage :
172
Lastpage :
177
Abstract :
This paper discusses the situated perception of a partner robot interacting with a human. The situated perception is based on the mutual cognitive environment discussed in relevance theory. It is very important to share actual environment in order to realize natural communication, because the meaning of natural language and gesture can be translated according to the surrounding environment. Next, we discuss the role of imitation based on gesture, and explain the method for imitative behavior generation. Finally, we show experimental results of situated perception, imitative behavior generation, and behavior coordination through interaction with a human.
Keywords :
fuzzy neural nets; learning (artificial intelligence); robots; behavior coordination; evolutionary computation; human-robot interaction; imitative behavior generation; imitative learning; neuro-fuzzy computing; partner robot; relevance theory; situated perception; spiking neural networks; Charge coupled devices; Cognitive robotics; Human robot interaction; Humanoid robots; Mobile robots; Natural languages; Robot kinematics; Robot sensing systems; Robot vision systems; Tactile sensors;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Advanced Robotics and its Social Impacts, 2005. IEEE Workshop on
Print_ISBN :
0-7803-8947-6
Type :
conf
DOI :
10.1109/ARSO.2005.1511646
Filename :
1511646
Link To Document :
بازگشت