Current state-of-the-art Anomaly Detection (AD) methods exploit the powerful representations yielded by large-scale ImageNet training. However, catastrophic forgetting prevents the successful fine-tuning of pre-trained representations on new datasets in the semi/unsupervised setting, and representations are therefore commonly fixed. In our work, we propose a new method to fine-tune learned representations for AD in a transfer learning setting. Based on the linkage between generative and discriminative modeling, we induce a multivariate Gaussian distribution for the normal class, and use the Mahalanobis distance of normal images to the distribution as training objective. We additionally propose to use augmentations commonly employed for vicinal risk minimization in a validation scheme to detect onset of catastrophic forgetting. Extensive evaluations on the public MVTec AD dataset reveal that a new state of the art is achieved by our method in the AD task while simultaneously achieving AS performance comparable to prior state of the art. Further, ablation studies demonstrate the importance of the induced Gaussian distribution as well as the robustness of the proposed fine-tuning scheme with respect to the choice of augmentations.
翻译:目前最先进的异常探测(AD)方法利用了大规模图像网络培训产生的强大表现方式。然而,灾难性的遗忘妨碍了在半/无人监督的环境中成功微调关于新数据集的预培训前展示方式,因此,这种展示方式是普遍的固定的。在我们的工作中,我们提出了一个新方法来微调在转让学习环境中所学的反倾销演示方式。根据基因化和歧视性模型之间的联系,我们为正常阶层引入了多种变式高斯分布,并将正常图像的距离马哈拉诺比斯作为分发的培训目标。我们还建议在验证计划中使用通常用于尽量减少振动风险的增强手段,以发现灾难性遗忘的开始。对公共MVTec AD数据集的广泛评估表明,我们通过在适应任务中的方法实现了新状态,同时实现了与以往的艺术状态相类似的AS性。此外,对通货膨胀的研究表明,诱导高斯分布的重要性以及拟议在增强能力选择方面的微调计划是稳健的。