DocumentCode :
2174077
Title :
The catchment feature model for multimodal language analysis
Author :
Quek, Francis
Author_Institution :
Vision Interfaces & Syst. Lab, Wright State Univ., Dayton, OH, USA
fYear :
2003
fDate :
13-16 Oct. 2003
Firstpage :
540
Abstract :
The catchment feature model (CFM) addresses two questions in multimodal interaction: how do we bridge video and audio processing with the realities of human multimodal communication, and how information from the different modes may be fused. We discuss the need for our model, motivate the CFM from psycholinguistic research, and present the model. In contrast to ´whole gesture´ recognition, the CFM applies a feature decomposition approach that facilitates cross-modal fusion at the level of discourse planning and conceptualization. We present our experimental framework for CFM-based research, and cite three concrete examples of catchment features (CF), and propose new directions of multimodal research based on the model.
Keywords :
audio signal processing; computer vision; feature extraction; gesture recognition; linguistics; natural languages; video signal processing; audio processing; catchment feature model; computer vision; cross-modal fusion; feature decomposition approach; gesture recognition; human multimodal communication; multimodal language analysis; video processing; Biological system modeling; Bridges; Concrete; Eyebrows; Face; Feedback; Humans; Psychology; Shape control; Speech;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Computer Vision, 2003. Proceedings. Ninth IEEE International Conference on
Conference_Location :
Nice, France
Print_ISBN :
0-7695-1950-4
Type :
conf
DOI :
10.1109/ICCV.2003.1238394
Filename :
1238394
Link To Document :
بازگشت