We present an explicit-grid based method for efficiently reconstructing streaming radiance fields for novel view synthesis of real world dynamic scenes. Instead of training a single model that combines all the frames, we formulate the dynamic modeling problem with an incremental learning paradigm in which per-frame model difference is trained to complement the adaption of a base model on the current frame. By exploiting the simple yet effective tuning strategy with narrow bands, the proposed method realizes a feasible framework for handling video sequences on-the-fly with high training efficiency. The storage overhead induced by using explicit grid representations can be significantly reduced through the use of model difference based compression. We also introduce an efficient strategy to further accelerate model optimization for each frame. Experiments on challenging video sequences demonstrate that our approach is capable of achieving a training speed of 15 seconds per-frame with competitive rendering quality, which attains $1000 \times$ speedup over the state-of-the-art implicit methods. Code is available at https://github.com/AlgoHunt/StreamRF.
翻译:我们提出了一个清晰的网络化方法,用于高效重建流光场,以进行真实世界动态场景的新视角合成。我们不但没有培训一个将所有框架结合起来的单一模型,而是用一个渐进式学习模式来制定动态模型问题,其中每个框架模型的差异经过培训,以补充当前框架基准模型的调整。通过利用狭小带的简单而有效的调控战略,拟议方法实现了在飞行时处理视频序列的可行框架,培训效率很高。通过使用基于压缩的模型差异模型,可以大大减少使用明确的网格表示方式诱发的存储间接费用。我们还引入了进一步加快每个框架模型优化的高效战略。关于富有挑战性的视频序列的实验表明,我们的方法能够达到每框架15秒钟的培训速度,具有竞争性的制作质量,在州艺术隐含方法上达到1,000美元的速度。代码可在https://github.com/AlgoHunt/StreamRF查阅。