Title :
Dynamic gestures as an input device for directing a mobile platform
Author :
Ehreumann, M. ; Lutticke, T. ; Dillmann, R.
Author_Institution :
Inst. for Process Control & Robotics, Karlsruhe Univ., Germany
Abstract :
Giving an advice to a mobile robot still requires classical user interfaces. A more intuitive way of commanding can be provided by verbal or gesture commands. In this article, we present new approaches and enhancements for established methods that are in use in our laboratory. Our aim is to direct a robot with simple dynamic gestures. We focus on visual gesture recognition. Based on skin color segmentation algorithms for tracking the user´s hand, hidden Markov models are used for gesture type recognition. The filters applied to the recorded trajectory strongly compress the input data. They also mark start and end point of a possible gesture. The hidden Markov models have been enhanced by a threshold model in order to wipe out insignificant movements. Pre-classification of the reference gestures serves for keeping computational effort low.
Keywords :
gesture recognition; hidden Markov models; image colour analysis; image segmentation; learning systems; mobile robots; robot programming; robot vision; dynamic gestures; hidden Markov models; learning system; mobile robot; robot programming; robot vision; skin color segmentation; user hand tracking; visual gesture recognition; Data gloves; Fingers; Hidden Markov models; Humans; Magnetic sensors; Mobile robots; Process control; Skin; Tracking; User interfaces;
Conference_Titel :
Robotics and Automation, 2001. Proceedings 2001 ICRA. IEEE International Conference on
Print_ISBN :
0-7803-6576-3
DOI :
10.1109/ROBOT.2001.933014