DocumentCode
2471702
Title
Vowel recognition from continuous articulatory movements for speaker-dependent applications
Author
Wang, Jun ; Green, Jordan R. ; Samal, Ashok ; Carrell, Tom D.
Author_Institution
Dept. of Comput. Sci. & Eng., Univ. of Nebraska - Lincoln, Lincoln, NE, USA
fYear
2010
fDate
13-15 Dec. 2010
Firstpage
1
Lastpage
7
Abstract
A novel approach was developed to recognize vowels from continuous tongue and lip movements. Vowels were classified based on movement patterns (rather than on derived articulatory features, e.g., lip opening) using a machine learning approach. Recognition accuracy on a single-speaker dataset was 94.02% with a very short latency. Recognition accuracy was better for high vowels than for low vowels. This finding parallels previous empirical findings on tongue movements during vowels. The recognition algorithm was then used to drive an articulation-to-acoustics synthesizer. The synthesizer recognizes vowels from continuous input stream of tongue and lip movements and plays the corresponding sound samples in near real-time.
Keywords
image recognition; learning (artificial intelligence); speech recognition; articulation-to-acoustics synthesizer; continuous articulatory movements; lip movements; machine learning; tongue movements; vowel recognition; Accuracy; Classification algorithms; Delay; Speech; Speech recognition; Tongue; Training; articulation; machine learning; recognition; support vector machine;
fLanguage
English
Publisher
ieee
Conference_Titel
Signal Processing and Communication Systems (ICSPCS), 2010 4th International Conference on
Conference_Location
Gold Coast, QLD
Print_ISBN
978-1-4244-7908-5
Electronic_ISBN
978-1-4244-7906-1
Type
conf
DOI
10.1109/ICSPCS.2010.5709716
Filename
5709716
Link To Document