Cost-sensitive classification is critical in applications where misclassification errors widely vary in cost. However, over-parameterization poses fundamental challenges to the cost-sensitive modeling of deep neural networks (DNNs). The ability of a DNN to fully interpolate a training dataset can render a DNN, evaluated purely on the training set, ineffective in distinguishing a cost-sensitive solution from its overall accuracy maximization counterpart. This necessitates rethinking cost-sensitive classification in DNNs. To address this challenge, this paper proposes a cost-sensitive adversarial data augmentation (CSADA) framework to make over-parameterized models cost-sensitive. The overarching idea is to generate targeted adversarial examples that push the decision boundary in cost-aware directions. These targeted adversarial samples are generated by maximizing the probability of critical misclassifications and used to train a model with more conservative decisions on costly pairs. Experiments on well-known datasets and a pharmacy medication image (PMI) dataset made publicly available show that our method can effectively minimize the overall cost and reduce critical errors, while achieving comparable performance in terms of overall accuracy.
翻译:在错误分类错误在成本上差异很大的应用中,成本敏感分类至关重要。然而,过度参数化对深神经网络的成本敏感模型构成根本性挑战。DNN完全内插培训数据集的能力可以使DNN完全内插,只对培训数据集进行评价,在区分成本敏感解决方案与总体准确度最大化对应方之间没有效果。这就要求重新思考DNN的成本敏感分类。为了应对这一挑战,本文件建议建立一个成本敏感的对抗数据增强框架,使多参数化模型具有成本敏感性。总体构想是产生有针对性的对抗性实例,在成本意识方向上推动决定界限。这些有针对性的对抗性样本的产生,是因为尽可能扩大关键分类的概率,并用于培训一个模型,对成本较高的配对作出更为保守的决定。关于众所周知的数据集和药用药物图像的实验公开显示,我们的方法可以有效地最大限度地降低总体成本,减少关键错误,同时在总体准确性方面实现可比较的业绩。