Current two-stage TTS framework typically integrates an acoustic model with a vocoder -- the acoustic model predicts a low resolution intermediate representation such as Mel-spectrum while the vocoder generates waveform from the intermediate representation. Although the intermediate representation is served as a bridge, there still exists critical mismatch between the acoustic model and the vocoder as they are commonly separately learned and work on different distributions of representation, leading to inevitable artifacts in the synthesized speech. In this work, different from using pre-designed intermediate representation in most previous studies, we propose to use VAE combining with GAN to learn a latent representation directly from speech and then utilize a flow-based acoustic model to model the distribution of the latent representation from text. In this way, the mismatch problem is migrated as the two stages work on the same distribution. Results demonstrate that the flow-based acoustic model can exactly model the distribution of our learned speech representation and the proposed TTS framework, namely Glow-WaveGAN, can produce high fidelity speech outperforming the state-of-the-art GAN-based model.
翻译:目前的双阶段 TTS 框架通常将声学模型与电码器(vocoder)结合 -- -- 声学模型预测低分辨率中间代表,如Mel-spectrum,而vocoder则从中间代表制产生波形。虽然中间代表制起到桥梁的作用,但声学模型和电码器之间仍然存在着严重的不匹配,因为它们通常是单独学习的,并致力于不同分布的表达式,从而导致在合成演讲中出现不可避免的人工制品。在这项工作中,与大多数先前的研究中使用预先设计的中间代表制不同的是,我们提议使用VAE,与GAN相结合的VAE,直接从语言中学习潜在代表制,然后使用流基声学模型来模拟文本潜在代表制的分布。这样,不匹配问题作为同一分布的两阶段工作而迁移。结果表明,流基声音模型可以准确地模拟我们所学语言代表式的分布和拟议的TTS-TS框架(即Glow-WaveGAN),能够产生高忠诚性演讲优于以状态为基础的GAN模型。