When deployed in practical applications, computer vision systems will encounter numerous unexpected images (\emph{{i.e.}}, out-of-distribution data). Due to the potentially raised safety risks, these aforementioned unseen data should be carefully identified and handled. Generally, existing approaches in dealing with out-of-distribution (OOD) detection mainly focus on the statistical difference between the features of OOD and in-distribution (ID) data extracted by the classifiers. Although many of these schemes have brought considerable performance improvements, reducing the false positive rate (FPR) when processing open-set images, they necessarily lack reliable theoretical analysis and generalization guarantees. Unlike the observed ways, in this paper, we investigate the OOD detection problem based on the Bayes rule and present a convincing description of the reason for failures encountered by conventional classifiers. Concretely, our analysis reveals that refining the probability distribution yielded by the vanilla neural networks is necessary for OOD detection, alleviating the issues of assigning high confidence to OOD data. To achieve this effortlessly, we propose an ultra-effective method to generate near-realistic outlier supervision. Extensive experiments on large-scale benchmarks reveal that our proposed \texttt{BayesAug} significantly reduces the FPR95 over 12.50\% compared with the previous schemes, boosting the reliability of machine learning systems. The code will be made publicly available.
翻译:在实际应用中,计算机视觉系统将遇到许多出乎意料的图像(hemph ⁇ i.e. ⁇ ),计算机视觉系统将遇到许多出乎意料的图像(gemphé ⁇ i.e. ⁇ ),由于潜在的安全风险,应当仔细查明和处理上述上述无形数据。一般而言,处理分配外(OOOD)的现有探测方法主要侧重于分类者所提取的OOOD和分配内(ID)数据之间的统计差异。虽然其中许多方案带来了相当大的性能改进,降低了处理开放图像时的假正率(FPR),但它们必然缺乏可靠的理论分析和一般化保障。与本文中观察到的方法不同,我们根据Bayes规则调查OOOD探测问题,并对常规分类者遇到的失败原因作出令人信服的说明。具体地说,我们的分析表明,改进香草神经网络产生的概率分布对于OOD检测是必要的,减轻对OD数据高度信任的问题。为了不费力地实现这一点,我们提议了一种极有效的方法来产生近现实的外部监督。我们观察到的方法是,我们所观察到的基于Bay规则的大规模实验,大规模地试验将大大地显示我们所拟议的推反的F-rasmay的系统。