Multilingual speech recognition with neural networks is often implemented with batch-learning, when all of the languages are available before training. An ability to add new languages after the prior training sessions can be economically beneficial, but the main challenge is catastrophic forgetting. In this work, we combine the qualities of weight factorization, transfer learning and Elastic Weight Consolidation in order to counter catastrophic forgetting and facilitate learning new languages quickly. Such combination allowed us to eliminate catastrophic forgetting while still achieving performance for the new languages comparable with having all languages at once, in experiments of learning from an initial 10 languages to achieve 27 languages
翻译:通过神经网络的多语言语音识别,往往通过批量学习来实施神经网络的多语言语言识别,当时所有语言在培训前都可以使用,在培训前的培训课程后增加新语言的能力在经济上可能有益,但主要挑战是灾难性的遗忘。在这项工作中,我们结合了重力因素化、转移学习和弹性体重整合等质量,以对抗灾难性遗忘,促进迅速学习新语言。这种结合使我们能够消除灾难性的遗忘,同时仍然能够实现与同时使用所有语言相比的新语言的功能,从最初的10种语言学习,以达到27种语言。