Recent work on speech self-supervised learning (speech SSL) demonstrated the benefits of scale in learning rich and transferable representations for Automatic Speech Recognition (ASR) with limited parallel data. It is then natural to investigate the existence of sparse and transferrable subnetworks in pre-trained speech SSL models that can achieve even better low-resource ASR performance. However, directly applying widely adopted pruning methods such as the Lottery Ticket Hypothesis (LTH) is suboptimal in the computational cost needed. Moreover, contrary to what LTH predicts, the discovered subnetworks yield minimal performance gain compared to the original dense network. In this work, we propose Prune-Adjust- Re-Prune (PARP), which discovers and finetunes subnetworks for much better ASR performance, while only requiring a single downstream finetuning run. PARP is inspired by our surprising observation that subnetworks pruned for pre-training tasks only needed to be slightly adjusted to achieve a sizeable performance boost in downstream ASR tasks. Extensive experiments on low-resource English and multi-lingual ASR show (1) sparse subnetworks exist in pre-trained speech SSL, and (2) the computational advantage and performance gain of PARP over baseline pruning methods. On the 10min Librispeech split without LM decoding, PARP discovers subnetworks from wav2vec 2.0 with an absolute 10.9%/12.6% WER decrease compared to the full model. We demonstrate PARP mitigates performance degradation in cross-lingual mask transfer, and investigate the possibility of discovering a single subnetwork for 10 spoken languages in one run.
翻译:最近关于语言自我监督学习的工作(Speech SSL)展示了在学习丰富和可转移的自动语音识别演示(ASR)时,以有限的平行数据学习丰富和可转移的演示(ASR)的规模的好处。然后自然地调查在经过训练的语音 SSL模型中存在的稀少和可转移的子网络,这些网络可以取得更好的 ASR 性能。然而,直接应用广泛采用的修剪方法,如Lottery Ticket Hypothes(LTH),在所需计算费用中不尽理想。此外,与LTH预测的相反,所发现的子网络与原始密度网络相比,取得了微小的绩效增益。在这项工作中,我们提议对低资源Adjust-Adjust-Reprint-Recrune (PARP) 进行广泛的实验,发现和微小的子网络为更好的 ASR 10-RP OV 的升级,我们从低资源和多语言的语音 Outal-LARD 的亚性能在SR 10 Ral 上展示了SB 10 的SLM 的升级的升级的一小的成绩分析方法。