DocumentCode
2174077
Title
The catchment feature model for multimodal language analysis
Author
Quek, Francis
Author_Institution
Vision Interfaces & Syst. Lab, Wright State Univ., Dayton, OH, USA
fYear
2003
fDate
13-16 Oct. 2003
Firstpage
540
Abstract
The catchment feature model (CFM) addresses two questions in multimodal interaction: how do we bridge video and audio processing with the realities of human multimodal communication, and how information from the different modes may be fused. We discuss the need for our model, motivate the CFM from psycholinguistic research, and present the model. In contrast to ´whole gesture´ recognition, the CFM applies a feature decomposition approach that facilitates cross-modal fusion at the level of discourse planning and conceptualization. We present our experimental framework for CFM-based research, and cite three concrete examples of catchment features (CF), and propose new directions of multimodal research based on the model.
Keywords
audio signal processing; computer vision; feature extraction; gesture recognition; linguistics; natural languages; video signal processing; audio processing; catchment feature model; computer vision; cross-modal fusion; feature decomposition approach; gesture recognition; human multimodal communication; multimodal language analysis; video processing; Biological system modeling; Bridges; Concrete; Eyebrows; Face; Feedback; Humans; Psychology; Shape control; Speech;
fLanguage
English
Publisher
ieee
Conference_Titel
Computer Vision, 2003. Proceedings. Ninth IEEE International Conference on
Conference_Location
Nice, France
Print_ISBN
0-7695-1950-4
Type
conf
DOI
10.1109/ICCV.2003.1238394
Filename
1238394
Link To Document