In one-class-learning tasks, only the normal case (foreground) can be modeled with data, whereas the variation of all possible anomalies is too erratic to be described by samples. Thus, due to the lack of representative data, the wide-spread discriminative approaches cannot cover such learning tasks, and rather generative models, which attempt to learn the input density of the foreground, are used. However, generative models suffer from a large input dimensionality (as in images) and are typically inefficient learners. We propose to learn the data distribution of the foreground more efficiently with a multi-hypotheses autoencoder. Moreover, the model is criticized by a discriminator, which prevents artificial data modes not supported by data, and enforces diversity across hypotheses. Our multiple-hypothesesbased anomaly detection framework allows the reliable identification of out-of-distribution samples. For anomaly detection on CIFAR-10, it yields up to 3.9% points improvement over previously reported results. On a real anomaly detection task, the approach reduces the error of the baseline models from 6.8% to 1.5%.
翻译:在单级学习任务中,只有普通案例(前方)可以模拟数据,而所有可能的异常现象的变异性则过于不稳定,无法通过样本来描述。因此,由于缺乏代表性数据,广泛广泛的歧视性方法无法涵盖这种学习任务,而是采用了基因化模型,试图了解前方的输入密度,但是,基因化模型存在大量输入的维度(如图像),而且通常是效率低下的学习者。我们提议以多功能自动编码器来更有效地学习前方数据分布。此外,该模型受到歧视者的批评,因为歧视者防止了没有数据支持的人工数据模式,并实施了跨假体的多样性。我们基于多种假体的异常检测框架使得能够可靠地识别分配外样本。在CIFAR-10上,异常检测结果比以前报告的结果提高3.9%。在真正的异常检测任务中,该方法将基线模型的错误从6.8%降至1.5%。