We present a new method for lightweight novel-view synthesis that generalizes to an arbitrary forward-facing scene. Recent approaches are computationally expensive, require per-scene optimization, or produce a memory-expensive representation. We start by representing the scene with a set of fronto-parallel semitransparent planes and afterward convert them to deformable layers in an end-to-end manner. Additionally, we employ a feed-forward refinement procedure that corrects the estimated representation by aggregating information from input views. Our method does not require fine-tuning when a new scene is processed and can handle an arbitrary number of views without restrictions. Experimental results show that our approach surpasses recent models in terms of common metrics and human evaluation, with the noticeable advantage in inference speed and compactness of the inferred layered geometry, see https://samsunglabs.github.io/MLI
翻译:我们提出了一种新的轻量级新观点合成方法,该方法可概括为任意的前瞻性场景。最近的方法在计算上成本高昂,需要每片优化,或产生一个内存成本代表。我们首先以一组前面和平行半透明式半透明式飞机代表现场,然后以端到端的方式将其转换为可变形层。此外,我们采用一个向后改进程序,通过汇总输入视图中的信息来纠正估计的表示方式。我们的方法不需要在新场景处理时进行微调,并且可以不受限制地处理任意数量的观察。实验结果显示,我们的方法在通用指标和人类评估方面超过了最近的模型,在推断的多层几层几何测量的推论速度和紧凑性方面有明显的优势,见https://samsunglabs.github.io/MLI。