DocumentCode :
54103
Title :
Voice Conversion Using Deep Neural Networks With Layer-Wise Generative Training
Author :
Ling-Hui Chen ; Zhen-Hua Ling ; Li-Juan Liu ; Li-Rong Dai
Author_Institution :
Nat. Eng. Lab. of Speech & Language Inf. Process., Univ. of Sci. & Technol. of China, Hefei, China
Volume :
22
Issue :
12
fYear :
2014
fDate :
Dec. 2014
Firstpage :
1859
Lastpage :
1872
Abstract :
This paper presents a new spectral envelope conversion method using deep neural networks (DNNs). The conventional joint density Gaussian mixture model (JDGMM) based spectral conversion methods perform stably and effectively. However, the speech generated by these methods suffer severe quality degradation due to the following two factors: 1) inadequacy of JDGMM in modeling the distribution of spectral features as well as the non-linear mapping relationship between the source and target speakers, 2) spectral detail loss caused by the use of high-level spectral features such as mel-cepstra. Previously, we have proposed to use the mixture of restricted Boltzmann machines (MoRBM) and the mixture of Gaussian bidirectional associative memories (MoGBAM) to cope with these problems. In this paper, we propose to use a DNN to construct a global non-linear mapping relationship between the spectral envelopes of two speakers. The proposed DNN is generatively trained by cascading two RBMs, which model the distributions of spectral envelopes of source and target speakers respectively, using a Bernoulli BAM (BBAM). Therefore, the proposed training method takes the advantage of the strong modeling ability of RBMs in modeling the distribution of spectral envelopes and the superiority of BAMs in deriving the conditional distributions for conversion. Careful comparisons and analysis among the proposed method and some conventional methods are presented in this paper. The subjective results show that the proposed method can significantly improve the performance in terms of both similarity and naturalness compared to conventional methods.
Keywords :
Boltzmann machines; speech synthesis; Bernoulli BAM; DNN; JDGMM; MoGBAM; MoRBM; deep neural networks; joint density Gaussian mixture model; layer-wise generative training; mel-cepstra; mixture of Gaussian bidirectional associative memories; mixture of restricted Boltzmann machines; quality degradation; spectral conversion methods; voice conversion; Covariance matrices; Joints; Neural networks; Speech; Speech processing; Stochastic processes; Training; Bidirectional associative memory; Gaussian mixture model; deep neural network; restricted Boltzmann machine; spectral envelope conversion; voice conversion;
fLanguage :
English
Journal_Title :
Audio, Speech, and Language Processing, IEEE/ACM Transactions on
Publisher :
ieee
ISSN :
2329-9290
Type :
jour
DOI :
10.1109/TASLP.2014.2353991
Filename :
6891242
Link To Document :
بازگشت