With the maturing of deep learning systems, trustworthiness is becoming increasingly important for model assessment. We understand trustworthiness as the combination of explainability and robustness. Generative classifiers (GCs) are a promising class of models that are said to naturally accomplish these qualities. However, this has mostly been demonstrated on simple datasets such as MNIST and CIFAR in the past. In this work, we firstly develop an architecture and training scheme that allows GCs to operate on a more relevant level of complexity for practical computer vision, namely the ImageNet challenge. Secondly, we demonstrate the immense potential of GCs for trustworthy image classification. Explainability and some aspects of robustness are vastly improved compared to feed-forward models, even when the GCs are just applied naively. While not all trustworthiness problems are solved completely, we observe that GCs are a highly promising basis for further algorithms and modifications. We release our trained model for download in the hope that it serves as a starting point for other generative classification tasks, in much the same way as pretrained ResNet architectures do for discriminative classification.
翻译:随着深层学习系统的成熟,可信赖性对模型评估越来越重要。我们理解可信任性是可解释性和稳健性的结合。创用分类(GCs)是一个很有希望的模型类别,据说自然地实现了这些品质。然而,这过去主要表现在诸如MNIST和CIFAR等简单数据集上。在这项工作中,我们首先开发了一个架构和培训计划,使GCs能够以更相关的复杂程度操作实用计算机愿景,即图像网的挑战。第二,我们展示了GCs在可信赖的图像分类方面的巨大潜力。可解释性和强健性的某些方面与饲料前方模型相比大为改善,即使GCs只是天真地应用。虽然并非所有可信任性问题都完全解决了,但我们认为,GCs是进一步算法和修改的极有希望的基础。我们推出经过培训的下载模型,希望它成为其他基因化分类任务的起点,其方式与事先培训的ResNet结构用于歧视性分类的大致相同。