Training robust Speech to Text (STT) system require "tens of thousand" of hours of data. Variability present in the dataset, in the form of unwanted nuisances (noise) and biases (accent, gender or age) is the reason for the need of large datasets to learn general representations, which is unfeasible for low resource languages. A recently proposed deep learning approach to remove these unwanted features, called adversarial forgetting, was able to produce better results on computer vision tasks. Motivated by this, in this paper, we study the effect of de-entangling the accent information from the input speech signal on training STT systems. To this end, we use an information bottleneck architecture based on adversarial forgetting. This training scheme aims to enforce the model to learn general accent invariant speech representations. The trained STT model is tested on two unseen accents in the common voice V1. The results are in favour of STT model trained using the adversarial forgetting scheme.
翻译:培训强力的文本演讲系统需要“十小时”的数据。 数据集中以不想要的干扰(噪音)和偏向(偏好、性别或年龄)的形式存在的差异,是需要大型数据集来学习通用表达方式的原因,对于低资源语言来说,这是不可行的。最近提出的消除这些不想要的特征的深层次学习方法,称为对抗式遗忘,能够在计算机的视觉任务上产生更好的结果。在本文中,我们研究了从培训STT系统输入的演讲信号中去除口音信息的效果。为此,我们使用基于对抗式遗忘的信息瓶颈结构。这一培训计划旨在实施模式,以学习通用语音V1中的通用口音。经过培训的STT模式在通用语音V1中用两种看不见的口音进行测试。结果有利于使用对抗性遗忘计划培训的STT模式。