Joint machine learning models that allow synthesizing and classifying data often offer uneven performance between those tasks or are unstable to train. In this work, we depart from a set of empirical observations that indicate the usefulness of internal representations built by contemporary deep diffusion-based generative models not only for generating but also predicting. We then propose to extend the vanilla diffusion model with a classifier that allows for stable joint end-to-end training with shared parameterization between those objectives. The resulting joint diffusion model outperforms recent state-of-the-art hybrid methods in terms of both classification and generation quality on all evaluated benchmarks. On top of our joint training approach, we present how we can directly benefit from shared generative and discriminative representations by introducing a method for visual counterfactual explanations.
翻译:联合机器学习模型允许合成和分类数据,但往往在这些任务之间提供不同的性能,或者难以稳定地训练。本文从一组实证观察出发,这些观察表明当代深度扩散生成模型构建的内部表示不仅有助于生成,还可以预测。然后,我们提出扩展基本扩散模型的方法,其中包含一个分类器,允许稳定的联合端到端训练以及这两个目标之间的共享参数化。结果,联合扩散模型在所有经过评估的基准测试中都优于最近的最先进的混合方法,无论是在分类还是生成质量方面。在我们的联合培训方法的基础上,我们介绍了一种方法,可以通过引入一种可视的反事实解释来直接从共享的生成和区分表示中受益。