This paper addresses the open set recognition (OSR) problem, where the goal is to correctly classify samples of known classes while detecting unknown samples to reject. In the OSR problem, "unknown" is assumed to have infinite possibilities because we have no knowledge about unknowns until they emerge. Intuitively, the more an OSR system explores the possibilities of unknowns, the more likely it is to detect unknowns. Thus, this paper proposes a novel synthetic unknown class learning method that generates unknown-like samples while maintaining diversity between the generated samples and learns these samples. In addition to this unknown sample generation process, knowledge distillation is introduced to provide room for learning synthetic unknowns. By learning the unknown-like samples and known samples in an alternating manner, the proposed method can not only experience diverse synthetic unknowns but also reduce overgeneralization with respect to known classes. Experiments on several benchmark datasets show that the proposed method significantly outperforms other state-of-the-art approaches. It is also shown that realistic unknown digits can be generated and learned via the proposed method after training on the MNIST dataset.
翻译:本文讨论了开放的识别( OSR) 问题, 目的是正确分类已知类别样本,同时检测未知样本,加以拒绝。 在 OSR 问题中, “ 未知” 被假定具有无限的可能性, 因为我们不知道未知物, 直到这些未知物出现。 直觉看, OSR 系统越是探索未知物的可能性, 发现未知物的可能性就越大。 因此, 本文提出一种新的合成未知类学习方法, 产生类似未知的样本, 同时保持生成样本之间的多样性, 并学习这些样本。 除了这一未知的样本生成过程外, 还引入了知识蒸馏, 为学习合成未知物提供空间。 通过以交替方式学习未知的样本和已知样本, 拟议的方法不仅可以体验不同的合成未知物, 还可以减少已知类别方面的超常化。 几个基准数据集的实验表明, 拟议的方法大大超越了其他状态的样本, 并学习这些样本。 此外, 也表明, 在对 NNIST 数据设置的培训后, 可以通过拟议方法生成和学习现实的未知数位。