Detecting out-of-distribution (OOD) inputs is critical for safely deploying deep learning models in an open-world setting. However, existing OOD detection solutions can be brittle in the open world, facing various types of adversarial OOD inputs. While methods leveraging auxiliary OOD data have emerged, our analysis on illuminative examples reveals a key insight that the majority of auxiliary OOD examples may not meaningfully improve or even hurt the decision boundary of the OOD detector, which is also observed in empirical results on real data. In this paper, we provide a theoretically motivated method, Adversarial Training with informative Outlier Mining (ATOM), which improves the robustness of OOD detection. We show that, by mining informative auxiliary OOD data, one can significantly improve OOD detection performance, and somewhat surprisingly, generalize to unseen adversarial attacks. ATOM achieves state-of-the-art performance under a broad family of classic and adversarial OOD evaluation tasks. For example, on the CIFAR-10 in-distribution dataset, ATOM reduces the FPR (at TPR 95%) by up to 57.99% under adversarial OOD inputs, surpassing the previous best baseline by a large margin.
翻译:在开放世界环境中安全部署深层学习模式的关键在于检测分配(OOOD)投入,但是,现有的OOD检测解决方案在开放世界中可能会变得很不稳定,面临各种对抗性OOD投入。虽然利用辅助OOOD数据的方法已经出现,但我们对具有启发性的例子的分析揭示出一个关键的洞察力,即大多数辅助OOOD实例可能不会有意义地改善甚至伤害OOOD探测器的决定界限,这在真实数据的经验结果中也观察到。在本文中,我们提供了一种有理论动机的方法,即信息化外部采矿(ATOM)的反向培训,提高了OOOD检测的可靠性。我们表明,通过开采信息化辅助OOOD数据,可以大大改进OOOD检测性能,并有点令人惊讶地将一般化为看不见的对抗性攻击。在传统和对抗性ODD评估任务的广泛系列下,ATO取得最新业绩。例如,在分配数据集的CIFAR-10中,ATOMM将FPR(在TR 95%上) 降为最大基线,比前在对抗性OD下的大部分投入超过5799%。