Existing methods for 3D-aware image synthesis largely depend on the 3D pose distribution pre-estimated on the training set. An inaccurate estimation may mislead the model into learning faulty geometry. This work proposes PoF3D that frees generative radiance fields from the requirements of 3D pose priors. We first equip the generator with an efficient pose learner, which is able to infer a pose from a latent code, to approximate the underlying true pose distribution automatically. We then assign the discriminator a task to learn pose distribution under the supervision of the generator and to differentiate real and synthesized images with the predicted pose as the condition. The pose-free generator and the pose-aware discriminator are jointly trained in an adversarial manner. Extensive results on a couple of datasets confirm that the performance of our approach, regarding both image quality and geometry quality, is on par with state of the art. To our best knowledge, PoF3D demonstrates the feasibility of learning high-quality 3D-aware image synthesis without using 3D pose priors for the first time.
翻译:现有的3D感知图像合成方法很大程度上依赖于在训练集上预估的3D姿态分布。不准确的估计可能会误导模型学习有误的几何结构。本文提出PoF3D,使生成辐射场摆脱了对3D姿态先验的要求。我们首先为生成器配备了一个高效的姿态学习器,能够从潜变量中推断出姿态,自动近似潜在的真实姿态分布。然后,我们给鉴别器分配一个任务,在生成器的监督下学习姿态分布,并以预测姿态为条件区分真实和生成的图像。无姿态的生成器和姿态感知的鉴别器在对抗训练中联合训练。对几个数据集的广泛实验表明,我们的方法在图像质量和几何质量方面的性能与最先进的方法相当。据我们所知,PoF3D首次证明了在不使用3D姿态先验的情况下学习高质量3D感知图像合成的可行性。