Generative models have shown great promise in synthesizing photorealistic 3D objects, but they require large amounts of training data. We introduce SinGRAF, a 3D-aware generative model that is trained with a few input images of a single scene. Once trained, SinGRAF generates different realizations of this 3D scene that preserve the appearance of the input while varying scene layout. For this purpose, we build on recent progress in 3D GAN architectures and introduce a novel progressive-scale patch discrimination approach during training. With several experiments, we demonstrate that the results produced by SinGRAF outperform the closest related works in both quality and diversity by a large margin.
翻译:生成模型在合成逼真的三维物体方面表现出非常大的潜力,但需要大量的训练数据。我们引入了SinGRAF,这是一种三维感知的生成模型,它使用少量单场景输入图片进行训练。一旦训练完成,SinGRAF就能够生成保留输入外观但场景布局不同的三维场景的不同实现。为此,我们基于最近在三维GAN架构方面取得的进展,并在训练时引入了一种新的渐进式尺度的修补区别方法。通过几个实验,我们证明了SinGRAF产生的结果在质量和多样性方面都大大优于其他最近相关的工作。