We study the problem of novel view synthesis of a scene comprised of 3D objects. We propose a simple yet effective approach that is neither continuous nor implicit, challenging recent trends on view synthesis. We demonstrate that although continuous radiance field representations have gained a lot of attention due to their expressive power, our simple approach obtains comparable or even better novel view reconstruction quality comparing with state-of-the-art baselines while increasing rendering speed by over 400x. Our model is trained in a category-agnostic manner and does not require scene-specific optimization. Therefore, it is able to generalize novel view synthesis to object categories not seen during training. In addition, we show that with our simple formulation, we can use view synthesis as a self-supervision signal for efficient learning of 3D geometry without explicit 3D supervision.
翻译:我们研究了由三维天体构成的场景的新视角合成问题。我们建议了一种既非连续又非隐含的简单而有效的方法,对近期的视觉合成趋势提出挑战。我们表明,尽管连续的光亮实地代表由于其表达力而得到了很多关注,但我们的简单方法获得了与最新水平基线相比较的可比甚至更好的新颖的重建质量,同时提高了400x的转化速度。我们的模型是经过分类的认知方式培训的,不需要对场景进行优化。因此,它能够将新视角合成归纳为培训期间看不到的物体类别。此外,我们表明,通过简单的表述,我们可以将合成视为一种自我监督的信号,在没有明确的三维监督的情况下有效学习三维几何方法。