Title :
Learning semantic components from subsymbolic multimodal perception
Author :
Mangin, Olivier ; Oudeyer, Pierre-Yves
Author_Institution :
Flowers Team, INRIA / ENSTA-Paristech, Paris, France
Abstract :
Perceptual systems often include sensors from several modalities. However, existing robots do not yet sufficiently discover patterns that are spread over the flow of multimodal data they receive. In this paper we present a framework that learns a dictionary of words from full spoken utterances, together with a set of gestures from human demonstrations and the semantic connection between words and gestures. We explain how to use a nonnegative matrix factorization algorithm to learn a dictionary of components that represent meaningful elements present in the multimodal perception, without providing the system with a symbolic representation of the semantics. We illustrate this framework by showing how a learner discovers word-like components from observation of gestures made by a human together with spoken descriptions of the gestures, and how it captures the semantic association between the two.
Keywords :
dictionaries; gesture recognition; learning (artificial intelligence); matrix decomposition; robots; speech recognition; dictionary; gestures; human demonstrations; multimodal data; nonnegative matrix factorization algorithm; pattern discovery; perceptual systems; semantic association; semantic component learning; semantic connection; spoken descriptions; spoken utterances; subsymbolic multimodal perception; symbolic representation; word-like components; Acoustics; Dictionaries; Histograms; Robots; Semantics; Vectors; Visualization;
Conference_Titel :
Development and Learning and Epigenetic Robotics (ICDL), 2013 IEEE Third Joint International Conference on
Conference_Location :
Osaka
DOI :
10.1109/DevLrn.2013.6652563