Despite advances in image classification methods, detecting the samples not belonging to the training classes is still a challenging problem. There has been a burst of interest in this subject recently, which is called Open-Set Recognition (OSR). In OSR, the goal is to achieve both the classification and detecting out-of-distribution (OOD) samples. Several ideas have been proposed to push the empirical result further through complicated techniques. We believe that such complication is indeed not necessary. To this end, we have shown that Maximum Softmax Probability (MSP), as the simplest baseline for OSR, applied on Vision Transformers (ViTs) as the base classifier that is trained with non-OOD augmentations can surprisingly outperform many recent methods. Non-OOD augmentations are the ones that do not alter the data distribution by much. Our results outperform state-of-the-art in CIFAR-10 datasets, and is also better than most of the current methods in SVHN and MNIST. We show that training augmentation has a significant effect on the performance of ViTs in the OSR tasks, and while they should produce significant diversity in the augmented samples, the generated sample OOD-ness must remain limited.
翻译:尽管在图像分类方法方面取得了进展,但发现不属于训练班的样品仍是一个棘手的问题,最近对这个问题的兴趣大增,这个题目被称为开放识别(OSR),在OSR中,目标是实现分类和探测分发之外的样品。提出了若干想法,以便通过复杂的技术进一步推进经验结果。我们认为,这种复杂情况确实没有必要。为此目的,我们已表明,作为OSR的最简单基准,对作为接受非OOOD扩增训练的基础分类师的OSR应用最大软体概率(MSP)是OSR的最简单基准,我们显示,在接受OVERT扩增训练的基本分类师(VITs)上应用的OVED扩增可以令人惊讶地超过最近的许多方法。非OD扩增是不会改变数据分布的许多方法。我们的结果超越了CIFAR-10数据集的先进状态,也比SVHN和MNIST中目前大多数方法要好。我们已表明,培训扩增对VT在OSR任务中的性能产生显著的影响,而它们必须提高样品的多样性。