Detecting whether examples belong to a given in-distribution or are Out-Of-Distribution (OOD) requires identifying features specific to the in-distribution. In the absence of labels, these features can be learned by self-supervised techniques under the generic assumption that the most abstract features are those which are statistically most over-represented in comparison to other distributions from the same domain. In this work, we show that self-distillation of the in-distribution training set together with contrasting against negative examples derived from shifting transformation of auxiliary data strongly improves OOD detection. We find that this improvement depends on how the negative samples are generated. In particular, we observe that by leveraging negative samples, which keep the statistics of low-level features while changing the high-level semantics, higher average detection performance is obtained. Furthermore, good negative sampling strategies can be identified from the sensitivity of the OOD detection score. The efficiency of our approach is demonstrated across a diverse range of OOD detection problems, setting new benchmarks for unsupervised OOD detection in the visual domain.
翻译:检测样本属于某一分布或流出分布(OOD)是否属于某个特定分布或流出分布(OOD),需要确定分布中的具体特征。在没有标签的情况下,这些特征可以通过自我监督技术来学习,其通用假设是,最抽象特征是统计上与同一领域其他分布相比在统计上最占多数的那些特征。在这项工作中,我们表明,自我提炼在分配中的培训,与从辅助数据转换变化中得出的负面实例相比较,极大地改善了OOOD的检测。我们发现,这一改进取决于负面样本是如何生成的。我们特别注意到,通过利用负面样本,在改变高层次的语义学的同时保留低水平特征的统计数据,可以取得更高的平均检测性能。此外,从OOD检测分的敏感度中可以发现良好的负面取样战略。我们的方法的效率体现在一系列不同的OD的检测问题中,为在视觉领域未受监督的 OODD检测设定了新的基准。