We introduce $\infty$-Diff, a generative diffusion model which directly operates on infinite resolution data. By randomly sampling subsets of coordinates during training and learning to denoise the content at those coordinates, a continuous function is learned that allows sampling at arbitrary resolutions. In contrast to other recent infinite resolution generative models, our approach operates directly on the raw data, not requiring latent vector compression for context, using hypernetworks, nor relying on discrete components. As such, our approach achieves significantly higher sample quality, as evidenced by lower FID scores, as well as being able to effectively scale to higher resolutions than the training data while retaining detail.
翻译:我们介绍了 $∞$-Diff,一种可以直接处理无限分辨率数据的生成式扩散模型。在训练过程中,我们通过随机采样坐标子集并学习去噪点内容,来学习一个连续的函数,可以在任意分辨率进行采样。与其他最近的无限分辨率生成模型不同的是,我们的方法直接处理原始数据,而不需要用超网络的潜在矢量压缩来获取上下文,也不依赖于离散组件。因此,我们的方法实现了显著更高的样本质量,如更低的FID分数,同时还能有效地扩展到比训练数据更高的分辨率,同时保留细节。