This paper describes the NPU system submitted to Spoofing Aware Speaker Verification Challenge 2022. We particularly focus on the \textit{backend ensemble} for speaker verification and spoofing countermeasure from three aspects. Firstly, besides simple concatenation, we propose circulant matrix transformation and stacking for speaker embeddings and countermeasure embeddings. With the stacking operation of newly-defined circulant embeddings, we almost explore all the possible interactions between speaker embeddings and countermeasure embeddings. Secondly, we attempt different convolution neural networks to selectively fuse the embeddings' salient regions into channels with convolution kernels. Finally, we design parallel attention in 1D convolution neural networks to learn the global correlation in channel dimensions as well as to learn the important parts in feature dimensions. Meanwhile, we embed squeeze-and-excitation attention in 2D convolutional neural networks to learn the global dependence among speaker embeddings and countermeasure embeddings. Experimental results demonstrate that all the above methods are effective. After fusion of four well-trained models enhanced by the mentioned methods, the best SASV-EER, SPF-EER and SV-EER we achieve are 0.559\%, 0.354\% and 0.857\% on the evaluation set respectively. Together with the above contributions, our submission system achieves the fifth place in this challenge.
翻译:本文描述了提交给2022年“潜意识演讲者核查挑战”的NPU系统。 我们特别侧重于从三个方面将演讲者核查和反制措施的“ Textit{ backend comple ” 专门用于对演讲者进行校验和反制反制措施。 首先,除了简单的连接外,我们提议循环矩阵转换,并堆叠供演讲者嵌入和反制嵌入。同时,随着新定义的螺旋嵌入和反制嵌入的堆叠作业,我们几乎探索了所有可能的演讲者嵌入和反制嵌入之间的相互作用。第二,我们尝试不同的进化神经网络有选择地将嵌入的突出区域与“革命核心”连接起来。最后,我们设计了1D演进神经网络的平行关注,以学习频道层面的全球相关性,并学习特征层面的重要部分。与此同时,我们把挤压和感动感应力嵌入的神经网络,以了解发言者嵌入和反制嵌入的全球依赖性。 实验结果表明,所有上述方法都是有效的。经过精心训练的4个模型在以上,S-58和S-ER-RS-BS-S-SA-SA-58 和M-B-B-BE-B-S-B-S-BE-BE-BE-BE-BE-S-S-S-S-S-S-SA-S-S-S-S-S-S-S-S-SA-S-SA-SA-SA-SA-SA-S-S-S-S-S-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-SA-