Novel view synthesis is a long-standing problem in machine learning and computer vision. Significant progress has recently been made in developing neural scene representations and rendering techniques that synthesize photorealistic images from arbitrary views. These representations, however, are extremely slow to train and often also slow to render. Inspired by neural variants of image-based rendering, we develop a new neural rendering approach with the goal of quickly learning a high-quality representation which can also be rendered in real-time. Our approach, MetaNLR++, accomplishes this by using a unique combination of a neural shape representation and 2D CNN-based image feature extraction, aggregation, and re-projection. To push representation convergence times down to minutes, we leverage meta learning to learn neural shape and image feature priors which accelerate training. The optimized shape and image features can then be extracted using traditional graphics techniques and rendered in real time. We show that MetaNLR++ achieves similar or better novel view synthesis results in a fraction of the time that competing methods require.
翻译:在机器学习和计算机视觉方面,一个长期存在的新观点合成是机器学习和计算机视觉方面的一个长期问题。最近,在开发神经场景演示和合成来自任意观点的光现实图像的技术方面取得了显著进展。然而,这些演示在培训方面极其缓慢,而且往往也缓慢。在基于图像的造影神经变异的启发下,我们开发了一种新的神经造影方法,目的是快速学习高质量的演示,这也可以实时实现。我们的方法MetaNLR++,通过使用神经形状代表的独特组合和基于2DCNN图像特征的提取、聚合和再投影来实现这一目标。为了将代表的趋同时间推到几分钟,我们利用元式学习神经形状和图像特征来加速培训。然后利用传统的图形技术来提取优化的形状和图像特征,并实时实现。我们显示MetaNLR++在竞争方法需要的一小部分时间里取得了类似或更好的新观点合成结果。