The idea of using phonological features instead of phonemes as input to sequence-to-sequence TTS has been recently proposed for zero-shot multilingual speech synthesis. This approach is useful for code-switching, as it facilitates the seamless uttering of foreign text embedded in a stream of native text. In our work, we train a language-agnostic multispeaker model conditioned on a set of phonologically derived features common across different languages, with the goal of achieving cross-lingual speaker adaptation. We first experiment with the effect of language phonological similarity on cross-lingual TTS of several source-target language combinations. Subsequently, we fine-tune the model with very limited data of a new speaker's voice in either a seen or an unseen language, and achieve synthetic speech of equal quality, while preserving the target speaker's identity. With as few as 32 and 8 utterances of target speaker data, we obtain high speaker similarity scores and naturalness comparable to the corresponding literature. In the extreme case of only 2 available adaptation utterances, we find that our model behaves as a few-shot learner, as the performance is similar in both the seen and unseen adaptation language scenarios.
翻译:使用声学特征而不是音频来作为音序到序列的 TTS 输入输入器的想法最近被提议用于零点的多语种语音合成。 这种方法对代码转换有用, 因为它有助于将本地文本流中所含的外国文本无缝地发音。 在我们的工作中, 我们训练一种语言的不可知性多语种模式, 以不同语言共有的声学衍生特征为条件, 目标是实现跨语种语言的调适。 我们首先试验语言声学相似性对多种源源- 目标语言组合的跨语言 TTS 的影响。 随后, 我们用一种可见或看不见的语言对新语者声音的非常有限的数据微调该模式, 并实现同等质量的合成语言表达, 同时保留目标演讲者的身份。 我们只获得与相应语言数据相近32和8个语种的语种特征, 我们获得与相应文学相似的高语种相似的评分和自然性。 在仅有2个可用调词的极端例子中, 我们发现我们的模型表现为几张语言的学习者, 其性在视觉中都能看到类似。