Benchmarking initiatives support the meaningful comparison of competing solutions to prominent problems in speech and language processing. Successive benchmarking evaluations typically reflect a progressive evolution from ideal lab conditions towards to those encountered in the wild. ASVspoof, the spoofing and deepfake detection initiative and challenge series, has followed the same trend. This article provides a summary of the ASVspoof 2021 challenge and the results of 37 participating teams. For the logical access task, results indicate that countermeasures solutions are robust to newly introduced encoding and transmission effects. Results for the physical access task indicate the potential to detect replay attacks in real, as opposed to simulated physical spaces, but a lack of robustness to variations between simulated and real acoustic environments. The DF task, new to the 2021 edition, targets solutions to the detection of manipulated, compressed speech data posted online. While detection solutions offer some resilience to compression effects, they lack generalization across different source datasets. In addition to a summary of the top-performing systems for each task, new analyses of influential data factors and results for hidden data subsets, the article includes a review of post-challenge results, an outline of the principal challenge limitations and a road-map for the future of ASVspoof. Link to the ASVspoof challenge and related resources: https://www.asvspoof.org/index2021.html
翻译:连续基准评估通常反映从理想的实验室条件向野生环境中遇到的场景的逐步演变。ASVSpooof, 潜伏和深假探测举措和挑战系列,也遵循了同样的趋势。这一条概述了2021年ASVpoof的挑战和37个参与小组的结果。关于逻辑访问任务,结果显示,反措施解决办法对新引入的编码和传输效应是强有力的。物理访问任务的结果表明,在真实而不是模拟物理空间中发现重现攻击的可能性,但缺乏对模拟和真实声学环境之间变化的强大性。DF任务,新到2021年版,目标是探测经操纵的压缩语音数据的解决办法。虽然检测办法提供了对压缩效应的一些弹性,但缺乏对不同来源数据集的概括性。除了概述每项任务的最高性系统,对有影响的数据因素和隐藏数据子集的新分析外,文章还包括对事后分析和模拟和真实声学环境之间变化的潜力的审查。DF任务,新到2021年版,旨在探测经调整后、压缩的语音/ASV相关挑战。