Title :
Speaker selection algorithm using audio and video information in a cluttered environment
Author :
Lim, Yoonseob ; Choi, JongSuk
Author_Institution :
Center for Cognitive Robot. Res., Korea Inst. of Sci. & Technol., Seoul
Abstract :
In this paper, data association method using audio and video data together to localize targets in a cluttered environment and detect who is speaking to the robot is presented. Particle filter is applied to find an optimal association between target and measurements efficiently. State variables are composed of position and speaking status. To update the speaking state, we first evaluate the upcoming sound signal based on the cross-correlation between microphones and then calculate the likelihood for audio information. Visual measurement is applied to find an optimal association between target and visual observations. The number of targets that robot should interact is updated based on the vision and audio information. Several experimental data were collected beforehand and simulated on computer to verify the performance of proposed data association method for speaker selection problem in a cluttered environment.
Keywords :
audio signal processing; clutter; microphones; particle filtering (numerical methods); robot vision; sensor fusion; speaker recognition; video signal processing; audio information; cluttered environment; data association method; microphones; particle filter; robot; speaker selection algorithm; video information; Acoustic noise; Cognitive robotics; Microphones; Particle filters; Robot sensing systems; Robot vision systems; Spatial resolution; Target tracking; Time measurement; Working environment noise;
Conference_Titel :
Emerging Technologies and Factory Automation, 2008. ETFA 2008. IEEE International Conference on
Conference_Location :
Hamburg
Print_ISBN :
978-1-4244-1505-2
Electronic_ISBN :
978-1-4244-1506-9
DOI :
10.1109/ETFA.2008.4638386