Recent work has shown that imperceptible perturbations can be applied to craft unlearnable examples (ULEs), i.e. images whose content cannot be used to improve a classifier during training. In this paper, we reveal the road that researchers should follow for understanding ULEs and improving ULEs as they were originally formulated (ULEOs). The paper makes four contributions. First, we show that ULEOs exploit color and, consequently, their effects can be mitigated by simple grayscale pre-filtering, without resorting to adversarial training. Second, we propose an extension to ULEOs, which is called ULEO-GrayAugs, that forces the generated ULEs away from channel-wise color perturbations by making use of grayscale knowledge and data augmentations during optimization. Third, we show that ULEOs generated using Multi-Layer Perceptrons (MLPs) are effective in the case of complex Convolutional Neural Network (CNN) classifiers, suggesting that CNNs suffer specific vulnerability to ULEs. Fourth, we demonstrate that when a classifier is trained on ULEOs, adversarial training will prevent a drop in accuracy measured both on clean images and on adversarial images. Taken together, our contributions represent a substantial advance in the state of art of unlearnable examples, but also reveal important characteristics of their behavior that must be better understood in order to achieve further improvements.
翻译:最近的工作表明,不可察觉的扰动可以应用于编造不可读的例子(ULE),即内容无法用来改进训练中的分类器的图像。在本文中,我们揭示了研究人员为了解ULE和按照最初的编制方式改进ULE(ULEOs)而应当遵循的道路。文件作出了四点贡献。首先,我们显示ULEOs利用颜色,因此其影响可以通过简单的灰度预先过滤来减轻,而不必进行对抗性培训。第二,我们建议扩展ULEO,即其内容不能用来改进训练期间的分类器。在本文件中,我们揭示了研究人员为了了解ULEO(ULE),应该遵循的道路是了解ULEO(ULE),因此,使用多Layer Perceptrons(MLPs)产生的ULEOs(MLPs)在复杂的进化神经神经网络(CNN)分类中是有效的,表明CNNIS会进一步受到ULE(ULE)的特殊脆弱性。第四,我们表明,当所测量的图像的精确度培训时,在测量的精确度上,也能够代表其准确性的准确性分析。