Recently, adversarial attack methods have been developed to challenge the robustness of machine learning models. However, mainstream evaluation criteria experience limitations, even yielding discrepancies among results under different settings. By examining various attack algorithms, including gradient-based and query-based attacks, we notice the lack of a consensus on a uniform standard for unbiased performance evaluation. Accordingly, we propose a Piece-wise Sampling Curving (PSC) toolkit to effectively address the aforementioned discrepancy, by generating a comprehensive comparison among adversaries in a given range. In addition, the PSC toolkit offers options for balancing the computational cost and evaluation effectiveness. Experimental results demonstrate our PSC toolkit presents comprehensive comparisons of attack algorithms, significantly reducing discrepancies in practice.
翻译:最近,开发了对抗性攻击方法,以挑战机器学习模式的稳健性;然而,主流评价标准存在局限性,甚至在不同环境下造成结果差异;通过审查各种攻击算法,包括梯度攻击和询问攻击,我们注意到在无偏见业绩评价的统一标准方面缺乏共识;因此,我们提议了一个简单抽样缩小工具包,以有效解决上述差异,在特定范围内对对手进行全面比较;此外,PSC工具包提供了平衡计算成本和评价效力的选项;实验结果显示,我们的PSC工具包对攻击算法进行了全面比较,大大缩小了实际差异。