• DocumentCode
    3018964
  • Title

    Transfer Learning in Sign language

  • Author

    Farhadi, Ali ; Forsyth, David ; White, Ryan

  • Author_Institution
    Univ. of Illinois at Urbana-Champaign, Urbana
  • fYear
    2007
  • fDate
    17-22 June 2007
  • Firstpage
    1
  • Lastpage
    8
  • Abstract
    We build word models for American Sign Language (ASL) that transfer between different signers and different aspects. This is advantageous because one could use large amounts of labelled avatar data in combination with a smaller amount of labelled human data to spot a large number of words in human data. Transfer learning is possible because we represent blocks of video with novel intermediate discriminative features based on splits of the data. By constructing the same splits in avatar and human data and clustering appropriately, our features are both discriminative and semantically similar: across signers similar features imply similar words. We demonstrate transfer learning in two scenarios: from avatar to a frontally viewed human signer and from an avatar to human signer in a 3/4 view.
  • Keywords
    avatars; character recognition; gesture recognition; pattern clustering; video signal processing; American Sign Language; avatar; human signer; pattern clustering; transfer learning; words; Avatars; Computer vision; Deafness; Handicapped aids; Hidden Markov models; Humans; Robustness; Spatial databases; Torso; Vocabulary;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Computer Vision and Pattern Recognition, 2007. CVPR '07. IEEE Conference on
  • Conference_Location
    Minneapolis, MN
  • ISSN
    1063-6919
  • Print_ISBN
    1-4244-1179-3
  • Electronic_ISBN
    1063-6919
  • Type

    conf

  • DOI
    10.1109/CVPR.2007.383346
  • Filename
    4270344