In automatic speech recognition (ASR) rescoring, the hypothesis with the fewest errors should be selected from the n-best list using a language model (LM). However, LMs are usually trained to maximize the likelihood of correct word sequences, not to detect ASR errors. We propose an ASR rescoring method for directly detecting errors with ELECTRA, which is originally a pre-training method for NLP tasks. ELECTRA is pre-trained to predict whether each word is replaced by BERT or not, which can simulate ASR error detection on large text corpora. To make this pre-training closer to ASR error detection, we further propose an extended version of ELECTRA called phone-attentive ELECTRA (P-ELECTRA). In the pre-training of P-ELECTRA, each word is replaced by a phone-to-word conversion model, which leverages phone information to generate acoustically similar words. Since our rescoring method is optimized for detecting errors, it can also be used for word-level confidence estimation. Experimental evaluations on the Librispeech and TED-LIUM2 corpora show that our rescoring method with ELECTRA is competitive with conventional rescoring methods with faster inference. ELECTRA also performs better in confidence estimation than BERT because it can learn to detect inappropriate words not only in fine-tuning but also in pre-training.
翻译:在自动语音识别(ASR)分解自动语音识别(ASR)中,使用语言模型(LM)从最优列表中选择最小差错的假设。然而,LMM通常会接受培训,以最大限度地提高正确字序列的可能性,而不是发现ASR错误。我们建议了与ELECTRA(ELECTRA)直接检测错误的ASR重新校准方法,ELECTRA最初是NLP任务的培训前方法。ELECTRA(ELECTRA)最初是使用电话对字转换模型,该模型利用电话信息生成声学上相似的词。由于我们的Recoring方法可以模拟大文本子公司对ASR的错误检测。为了使培训前更接近于ASR错误检测,我们进一步建议扩大ELECTRA(P-ELECTRA)称为电话识别强化字母序列序列的扩展版本。在P-ELECTRA(E-LTRA)前培训中,每个词都由电话对词转换模型取代,该模型利用电话信息产生声学上相似的词。由于我们的Recoring the develop develop listris and Excience LABCUA 方法,因此在LA RECUB-CUB-CS-CS-CUBS-S-S-S-CRABS-S-CUBS-S-S-S-S-CUDRA中也显示更高级方法中进行更佳的测试方法,因此在LB-S-S-S-S-S-CUD-CS-CUD-S-CS-CR-S-S-S-S-S-S-S-S-S-S-S-S-C-S-S-C-C-C-C-C-C-C-C-C-C-C-C-C-C-C-C-C-C-C-C-C-C-S-S-S-S-C-C-C-C-C-C-C-C-C-C-C-C-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-