• DocumentCode
    2174077
  • Title

    The catchment feature model for multimodal language analysis

  • Author

    Quek, Francis

  • Author_Institution
    Vision Interfaces & Syst. Lab, Wright State Univ., Dayton, OH, USA
  • fYear
    2003
  • fDate
    13-16 Oct. 2003
  • Firstpage
    540
  • Abstract
    The catchment feature model (CFM) addresses two questions in multimodal interaction: how do we bridge video and audio processing with the realities of human multimodal communication, and how information from the different modes may be fused. We discuss the need for our model, motivate the CFM from psycholinguistic research, and present the model. In contrast to ´whole gesture´ recognition, the CFM applies a feature decomposition approach that facilitates cross-modal fusion at the level of discourse planning and conceptualization. We present our experimental framework for CFM-based research, and cite three concrete examples of catchment features (CF), and propose new directions of multimodal research based on the model.
  • Keywords
    audio signal processing; computer vision; feature extraction; gesture recognition; linguistics; natural languages; video signal processing; audio processing; catchment feature model; computer vision; cross-modal fusion; feature decomposition approach; gesture recognition; human multimodal communication; multimodal language analysis; video processing; Biological system modeling; Bridges; Concrete; Eyebrows; Face; Feedback; Humans; Psychology; Shape control; Speech;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Computer Vision, 2003. Proceedings. Ninth IEEE International Conference on
  • Conference_Location
    Nice, France
  • Print_ISBN
    0-7695-1950-4
  • Type

    conf

  • DOI
    10.1109/ICCV.2003.1238394
  • Filename
    1238394