Title :
Acoustic to articulatory parameter mapping using an assembly of neural networks
Author :
Rahim, M.G. ; Keijn, W.B. ; Schroeter, J. ; Goodyear, C.C.
Author_Institution :
AT&T Bell Labs., Murray Hill, NJ, USA
Abstract :
The authors describe an efficient procedure for acoustic-to-articulatory parameter mapping using neural networks. An assembly of multilayer perceptrons, each designated to a specific region in the articulatory space, is used to map acoustic parameters of the speech into tract areas. The training of this model is executed in two stages; in the first stage a codebook of suitably normalized articulatory parameters is used and in the second stage real speech data are used to further improve the mapping. In general, acoustic-to-articulatory parameter mapping is nonunique; several vocal tract shapes can result in identical spectral envelopes. The model accommodates this ambiguity. During synthesis, neural networks are selected by dynamic programming using a criterion that ensures smoothly varying vocal tract shapes while maintaining a good spectral match
Keywords :
dynamic programming; neural nets; speech synthesis; acoustic-to-articulatory parameter mapping; articulatory space; dynamic programming; identical spectral envelopes; multilayer perceptrons; neural networks; real speech data; speech synthesis; tract areas; vocal tract shapes; Assembly; Filters; Multilayer perceptrons; Network synthesis; Neural networks; Shape; Smoothing methods; Speech analysis; Speech synthesis; Synthesizers;
Conference_Titel :
Acoustics, Speech, and Signal Processing, 1991. ICASSP-91., 1991 International Conference on
Conference_Location :
Toronto, Ont.
Print_ISBN :
0-7803-0003-3
DOI :
10.1109/ICASSP.1991.150382