It is well-known that deep learning models are vulnerable to adversarial examples. Existing studies of adversarial training have made great progress against this challenge. As a typical trait, they often assume that the class distribution is overall balanced. However, long-tail datasets are ubiquitous in a wide spectrum of applications, where the amount of head class instances is larger than the tail classes. Under such a scenario, AUC is a much more reasonable metric than accuracy since it is insensitive toward class distribution. Motivated by this, we present an early trial to explore adversarial training methods to optimize AUC. The main challenge lies in that the positive and negative examples are tightly coupled in the objective function. As a direct result, one cannot generate adversarial examples without a full scan of the dataset. To address this issue, based on a concavity regularization scheme, we reformulate the AUC optimization problem as a saddle point problem, where the objective becomes an instance-wise function. This leads to an end-to-end training protocol. Furthermore, we provide a convergence guarantee of the proposed algorithm. Our analysis differs from the existing studies since the algorithm is asked to generate adversarial examples by calculating the gradient of a min-max problem. Finally, the extensive experimental results show the performance and robustness of our algorithm in three long-tail datasets.
翻译:众所周知,深层次的学习模式很容易受到对抗性实例的影响。现有的对抗性培训研究在应对这一挑战方面已经取得了很大进展。作为典型特征,它们往往认为课堂分布总体上是平衡的。然而,长尾数据集在广泛的应用中是无处不在的,因为头类案例的数量大于尾类。在这种假设下,AUC是一个比准确性更合理的衡量标准,因为它对阶级分布不敏感。受此驱动,我们提出一个早期试验,探索对抗性培训方法以优化ACU。主要挑战在于,正反两方面的例子在目标功能中紧密结合。作为直接结果,一个人无法在不全面扫描数据集的情况下产生对抗性实例。为了解决这个问题,我们根据凝固性规范化机制,将AUC优化问题重新定位为一个支撑点问题,因为其目标变得不切实际功能。这导致一个端对端培训协议。此外,我们提出的算法的趋同性保证。我们的分析与现有的研究不同,因为现行算法与现有研究不同,因为要求对数据集进行全面扫描,最终通过精确的实验性模型来显示我们高水平。