DocumentCode :
3709411
Title :
Multimodal joint visual attention model for natural human-robot interaction in domestic environments
Author :
Joris Domhof;Aswin Chandarr;Maja Rudinac;Pieter Jonker
Author_Institution :
Delft Robotics Institute, TU, The Netherlands
fYear :
2015
fDate :
9/1/2015 12:00:00 AM
Firstpage :
2406
Lastpage :
2412
Abstract :
In this paper, we introduce a non-verbal multimodal joint visual attention model for human-robot interaction in household scenarios. Our model combines the bottom-up saliency and depth-based segmentation with the top-down cues such as pointing and gaze to detect the objects of interest according to the user. For generation of the top-down saliency maps, we have introduced novel methods for object saliency, based on the pointing direction as well as the gaze direction. For gaze estimation, a hybrid model has been introduced which automatically selects keypoint-based matching or back-projection based on the textureness of the object model. The combination of different cues ensures reliable object detection and interaction independent of the relative position between the user, robot and objects. Extensive experiments show good detection results in different interaction scenarios as well as in challenging environmental conditions.
Keywords :
"Robots","Visualization","Image color analysis","Three-dimensional displays","Computational modeling","Cameras","Estimation"
Publisher :
ieee
Conference_Titel :
Intelligent Robots and Systems (IROS), 2015 IEEE/RSJ International Conference on
Type :
conf
DOI :
10.1109/IROS.2015.7353703
Filename :
7353703
Link To Document :
بازگشت