• DocumentCode
    2416745
  • Title

    Rapid unsupervised adaptation using context independent phoneme model

  • Author

    Kobashikawa, S. ; Ogawa, Anna ; Yamaguchi, Yoshio ; Takahashi, Satoshi

  • Author_Institution
    NTT Cyber Space Labs., NTT Corp., Yokosuka, Japan
  • fYear
    2009
  • fDate
    25-28 May 2009
  • Firstpage
    209
  • Lastpage
    212
  • Abstract
    Users require rapid and highly accurate speech recognition systems. Accuracy could be improved by unsupervised adaptation as provided by CMLLR (Constrained Maximum Likelihood Linear Regression). CMLLR-based batch-type unsupervised adaptation estimates a single global transformation matrix by utilizing unsupervised labeling; unfortunately, it needs prior labeling and so is not rapid. Our proposed technique reduces the prior labeling time by using context independent phoneme models (monophones) and frame-by-frame statistics accumulation in unsupervised adaptation. The proposed technique further raises the accuracy by accumulating statistics with power and performing recognition with power after adaptation. Simulations using spontaneous speech show that the proposed technique reduced the total computational time of labeling and recognition by 52.2% while matching the recognition rate of the conventional unsupervised adaptation technique that uses context dependent phoneme models (triphones) statistics accumulation.
  • Keywords
    speech enhancement; speech recognition; statistics; context independent phoneme model; frame-by-frame statistics accumulation; monophones; rapid unsupervised adaptation; speech recognition systems; spontaneous speech; triphones; Acoustic beams; Automatic speech recognition; Computational modeling; Consumer electronics; Context modeling; Labeling; Laboratories; Maximum likelihood linear regression; Speech recognition; Statistics;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Consumer Electronics, 2009. ISCE '09. IEEE 13th International Symposium on
  • Conference_Location
    Kyoto
  • Print_ISBN
    978-1-4244-2975-2
  • Electronic_ISBN
    978-1-4244-2976-9
  • Type

    conf

  • DOI
    10.1109/ISCE.2009.5157000
  • Filename
    5157000