The advancement of generative radiance fields has pushed the boundary of 3D-aware image synthesis. Motivated by the observation that a 3D object should look realistic from multiple viewpoints, these methods introduce a multi-view constraint as regularization to learn valid 3D radiance fields from 2D images. Despite the progress, they often fall short of capturing accurate 3D shapes due to the shape-color ambiguity, limiting their applicability in downstream tasks. In this work, we address this ambiguity by proposing a novel shading-guided generative implicit model that is able to learn a starkly improved shape representation. Our key insight is that an accurate 3D shape should also yield a realistic rendering under different lighting conditions. This multi-lighting constraint is realized by modeling illumination explicitly and performing shading with various lighting conditions. Gradients are derived by feeding the synthesized images to a discriminator. To compensate for the additional computational burden of calculating surface normals, we further devise an efficient volume rendering strategy via surface tracking, reducing the training and inference time by 24% and 48%, respectively. Our experiments on multiple datasets show that the proposed approach achieves photorealistic 3D-aware image synthesis while capturing accurate underlying 3D shapes. We demonstrate improved performance of our approach on 3D shape reconstruction against existing methods, and show its applicability on image relighting. Our code will be released at https://github.com/XingangPan/ShadeGAN.
翻译:突变光亮场的进步拉动了 3D 图像合成 3D 的边界。 3D 对象应该从多个角度看现实, 这些方法引入了多视图限制, 将常规化从 2D 图像中学习有效的 3D 光亮场。 尽管取得了进步, 但由于形状颜色模糊, 它们往往无法捕捉准确的 3D 形状。 限制其在下游任务中的应用性能。 在这项工作中, 我们通过提出一个新的阴影引导的基因隐含模型来解决这一模糊性, 从而能够学习得到明显改进的形状表示。 我们的关键洞察力是, 准确的 3D 形状也应该在不同照明条件下产生现实的显示。 这种多光亮的制约是通过明确制作照明效果模型和用各种照明条件的阴影来实现的。 通过将合成的图像反馈到一个歧视者身上来产生。 为了补偿计算地表正常的额外计算负担, 我们通过地表跟踪, 将培训和推断时间分别减少 24 % 和 48 % 。 我们在多个数据- D 模型上进行的实验将显示我们当前3D 3D 的图像的模型的模拟方法显示我们现有的合成方法。