Traditionally anomaly detection (AD) is treated as an unsupervised problem utilizing only normal samples due to the intractability of characterizing everything that looks unlike the normal data. However, it has recently been found that unsupervised image anomaly detection can be drastically improved through the utilization of huge corpora of random images to represent anomalousness; a technique which is known as Outlier Exposure. In this paper we show that specialized AD learning methods seem actually superfluous and huge corpora of data expendable. For a common AD benchmark on ImageNet, standard classifiers and semi-supervised one-class methods trained to discern between normal samples and just a few random natural images are able to outperform the current state of the art in deep AD, and only one useful outlier sample is sufficient to perform competitively. We investigate this phenomenon and reveal that one-class methods are more robust towards the particular choice of training outliers. Furthermore, we find that a simple classifier based on representations from CLIP, a recent foundation model, achieves state-of-the-art results on CIFAR-10 and also outperforms all previous AD methods on ImageNet without any training samples (i.e., in a zero-shot setting).
翻译:传统上异常现象的检测(AD)被视为一个不受监督的问题,仅使用正常的样本,因为所有看起来与正常数据不同的样本都难以定性。然而,最近发现,通过使用巨大的随机图像团团体来代表异常现象,不经监督的图像异常现象的检测可以大大改善;一种称为外部暴露的技术。在本文中,我们表明,专门的AD学习方法实际上似乎是多余的和巨大的消耗数据体。对于一个通用的图像网自动评估基准,标准分类器和半监督的单级方法,经过培训,能够辨别正常样本和少数随机自然图像,能够超越深层ADD中目前艺术状态,只有一种有用的外部样本足以进行竞争性的运行。我们调查这一现象,并表明,单级方法对于培训外部人员的特定选择更为有力。此外,我们发现,基于CLIP(最近的基础模型)的演示,一个简单的分类器,可以实现CFAR-10的状态艺术结果,并且也是在未经任何培训的情况下,在图像网上的所有前AD-FAR-10和外形方法。