Title :
An active audition framework for auditory-driven HRI: Application to interactive robot dancing
Author :
Oliveira, Joao Lobato ; Ince, Gökhan ; Nakamura, Keisuke ; Nakadai, Kazuhiro ; Okuno, Hiroshi G. ; Reis, Luis Paulo ; Gouyon, Fabien
Author_Institution :
Artificial Intell. & Comput. Sci. Lab., FEUP, Porto, Portugal
Abstract :
In this paper we propose a general active audition framework for auditory-driven Human-Robot Interaction (HRI). The proposed framework simultaneously processes speech and music on-the-fly, integrates perceptual models for robot audition, and supports verbal and non-verbal interactive communication by means of (pro)active behaviors. To ensure a reliable interaction, on top of the framework a behavior decision mechanism based on active audition policies the robot´s actions according to the reliability of the acoustic signals for auditory processing. To validate the framework´s application to general auditory-driven HRI, we propose the implementation of an interactive robot dancing system. This system integrates three preprocessing robot audition modules: sound source localization, sound source separation, and ego noise suppression; two modules for auditory perception: live audio beat tracking and automatic speech recognition; and multi-modal behaviors for verbal and non-verbal interaction: music-driven dancing and speech-driven dialoguing. To fully assess the system, we set up experimental and interactive real-world scenarios with highly dynamic acoustic conditions, and defined a set of evaluation criteria. The experimental tests revealed accurate and robust beat tracking and speech recognition, and convincing dance beat-synchrony. The interactive sessions confirmed the fundamental role of the behavior decision mechanism for actively maintaining a robust and natural human-robot interaction.
Keywords :
acoustic generators; acoustic radiators; acoustic signal processing; human-robot interaction; interactive systems; mobile robots; music; noise abatement; source separation; speech processing; speech recognition; acoustic signals; active audition framework; auditory perception; auditory processing; auditory-driven HRI; auditory-driven human-robot interaction; automatic speech recognition; behavior decision mechanism; dance beat-synchrony; ego noise suppression; interactive robot dancing; interactive sessions; live audio beat tracking; multimodal behaviors; music on-the-fly processing; music-driven dancing; nonverbal interaction; nonverbal interactive communication; preprocessing robot audition module; proactive behaviors; robot audition; robust beat tracking; sound source localization; sound source separation; speech processing; speech-driven dialoguing; verbal interaction; verbal interactive communication; Acoustics; Noise; Reliability; Robot kinematics; Robot sensing systems; Speech;
Conference_Titel :
RO-MAN, 2012 IEEE
Conference_Location :
Paris
Print_ISBN :
978-1-4673-4604-7
Electronic_ISBN :
1944-9445
DOI :
10.1109/ROMAN.2012.6343892