Title :
Speaker selection and tracking in a cluttered environment with audio and visual information
Author :
Lim, Yoonseob ; Choi, JongSuk
Author_Institution :
Korea Inst. of Sci. & Technol., Seoul, South Korea
fDate :
8/1/2009 12:00:00 AM
Abstract :
Presented in this paper is a data association method using audio and visual data which localizes targets in a cluttered environment and detects who is speaking to a robot. A particle filter is applied to efficiently select the optimal association between the target and the measurements. State variables are composed of target positions and speaking states. To update the speaking state, we first evaluate the incoming sound signal based on cross-correlation and then calculate a likelihood from the audio information. The visual measurement is used to find an optimal association between the target and the observed objects. The number of targets that the robot should interact with is updated from the existence probabilities and associations. Experimental data were collected beforehand and simulated on a computer to verify the performance of the proposed method applied to the speaker selection problem in a cluttered environment. The algorithm was also implemented in a robotic system to demonstrate reliable interactions between the robot and speaking targets.
Keywords :
human-robot interaction; robot vision; sensor fusion; speaker recognition; audio information; cluttered environment; data association method; human-robot interaction; particle filter; robotic system; speaker selection; speaker tracking; visual information; Acoustic noise; Face recognition; Humans; Particle filters; Particle measurements; Radar tracking; Robot sensing systems; Spatial resolution; Target tracking; Working environment noise; Data Association; Human-Robot Interaction; Particle Filter; Speaker Localization;
Journal_Title :
Consumer Electronics, IEEE Transactions on
DOI :
10.1109/TCE.2009.5278030