Randomized Smoothing (RS) is a promising method for obtaining robustness certificates by evaluating a base model under noise. In this work, we: (i) theoretically motivate why ensembles are a particularly suitable choice as base models for RS, and (ii) empirically confirm this choice, obtaining state-of-the-art results in multiple settings. The key insight of our work is that the reduced variance of ensembles over the perturbations introduced in RS leads to significantly more consistent classifications for a given input. This, in turn, leads to substantially increased certifiable radii for samples close to the decision boundary. Additionally, we introduce key optimizations which enable an up to 55-fold decrease in sample complexity of RS, thus drastically reducing its computational overhead. Experimentally, we show that ensembles of only 3 to 10 classifiers consistently improve on their strongest constituting model with respect to their average certified radius (ACR) by 5% to 21% on both CIFAR10 and ImageNet, achieving a new state-of-the-art ACR of 0.86 and 1.11, respectively. We release all code and models required to reproduce our results upon publication.
翻译:在这项工作中,我们:(一) 理论上激励人们为什么集合是一个特别合适的选择,作为RS的基础模型,以及(二) 经验性地确认这一选择,在多个环境中获得最新工艺结果。我们工作的关键见解是,在RS引入的扰动中,组合差异的减少导致对特定输入的分类明显更加一致。这反过来又导致对靠近决定边界的样本的可认证辐射量大幅度增加。此外,我们引入关键优化,使RS的样本复杂性减少多达55倍,从而大幅度降低其计算间接费用。我们实验性地表明,只有3至10个分类者组成的组合,其最强的构成模型在CIFAR10和图像网络的平均认证半径(ACR)方面不断提高5%至21%,从而分别实现0.86和1.11的更新结果。我们发布了所有必要的代码和模型。