In this paper, we study the problem of consistency in the context of adversarial examples. Specifically, we tackle the following question: can surrogate losses still be used as a proxy for minimizing the $0/1$ loss in the presence of an adversary that alters the inputs at test-time? Different from the standard classification task, this question cannot be reduced to a point-wise minimization problem, and calibration needs not to be sufficient to ensure consistency. In this paper, we expose some pathological behaviors specific to the adversarial problem, and show that no convex surrogate loss can be consistent or calibrated in this context. It is therefore necessary to design another class of surrogate functions that can be used to solve the adversarial consistency issue. As a first step towards designing such a class, we identify sufficient and necessary conditions for a surrogate loss to be calibrated in both the adversarial and standard settings. Finally, we give some directions for building a class of losses that could be consistent in the adversarial framework.
翻译:在本文中,我们研究了在对抗性例子中的一致性问题。具体地说,我们处理的一个问题是:代用损失是否仍然可以用作在试验时改变投入的对手在场的情况下最大限度地减少0.1美元损失的代用手段?不同于标准的分类任务,这个问题不能降低到一个微小的最小化问题,而校准需要不足以确保一致性。在本文中,我们暴露了对抗性问题中特有的一些病理行为,并表明在此情况下无法对锥形代用损失进行一致或校准。因此,有必要设计另一类代用功能,用于解决对抗性一致性问题。作为设计这种分类的第一步,我们为代用损失在对抗性和标准环境下加以校准确定充分和必要的条件。最后,我们为建立在对抗性框架中可以保持一致的一类损失提供了一些指导。