Volume rendering using neural fields has shown great promise in capturing and synthesizing novel views of 3D scenes. However, this type of approach requires querying the volume network at multiple points along each viewing ray in order to render an image, resulting in very slow rendering times. In this paper, we present a method that overcomes this limitation by learning a direct mapping from camera rays to locations along the ray that are most likely to influence the pixel's final appearance. Using this approach we are able to render, train and fine-tune a volumetrically-rendered neural field model an order of magnitude faster than standard approaches. Unlike existing methods, our approach works with general volumes and can be trained end-to-end.
翻译:使用神经场域的音量转换在捕捉和综合3D场景的新观点方面显示了巨大的希望。 但是,这种方式要求在每个查看光线的多个点查询音量网络,以便形成图像,造成非常缓慢的交替时间。 在本文中,我们提出了一个方法,通过学习从相机射线到最有可能影响像素最后外观的射线线沿线位置的直接绘图,克服了这一限制。 使用这种方法,我们可以制造、培训和微调一个数量级的神经场模型,比标准方法要快。 与现有的方法不同,我们的方法是用一般的射线和训练端到端。