We present a Conformer-based end-to-end neural diarization (EEND) model that uses both acoustic input and features derived from an automatic speech recognition (ASR) model. Two categories of features are explored: features derived directly from ASR output (phones, position-in-word and word boundaries) and features derived from a lexical speaker change detection model, trained by fine-tuning a pretrained BERT model on the ASR output. Three modifications to the Conformer-based EEND architecture are proposed to incorporate the features. First, ASR features are concatenated with acoustic features. Second, we propose a new attention mechanism called contextualized self-attention that utilizes ASR features to build robust speaker representations. Finally, multi-task learning is used to train the model to minimize classification loss for the ASR features along with diarization loss. Experiments on the two-speaker English conversations of Switchboard+SRE data sets show that multi-task learning with position-in-word information is the most effective way of utilizing ASR features, reducing the diarization error rate (DER) by 20% relative to the baseline.
翻译:我们提出了一个基于后端至端神经二分化(EEND)模型,该模型使用声学输入和自动语音识别(ASR)模型的特征,探索了两类特征:直接来自ASR输出的特征(电话、文中位置和字的界限)和由词汇式扬声器变化检测模型产生的特征,该模型通过对ASR输出的预先培训的BERT模型进行微调培训,对基于源头的EEND结构进行三项修改,以纳入这些特征。首先,AEND的功能与声学特征相融合。第二,我们建议建立一个称为背景化自我注意的新关注机制,利用ASR的特征构建强有力的语音演示。最后,多任务学习用于培训模型,以尽量减少ASR特征的分类损失以及分解损失。对总机+SRE数据集的双声英语对话的实验显示,用位置-文字信息学习是使用ASR特征的最有效方法,将二分解误率比基线降低20%。