The success of the Neural Radiance Fields (NeRFs) for modeling and free-view rendering static objects has inspired numerous attempts on dynamic scenes. Current techniques that utilize neural rendering for facilitating free-view videos (FVVs) are restricted to either offline rendering or are capable of processing only brief sequences with minimal motion. In this paper, we present a novel technique, Residual Radiance Field or ReRF, as a highly compact neural representation to achieve real-time FVV rendering on long-duration dynamic scenes. ReRF explicitly models the residual information between adjacent timestamps in the spatial-temporal feature space, with a global coordinate-based tiny MLP as the feature decoder. Specifically, ReRF employs a compact motion grid along with a residual feature grid to exploit inter-frame feature similarities. We show such a strategy can handle large motions without sacrificing quality. We further present a sequential training scheme to maintain the smoothness and the sparsity of the motion/residual grids. Based on ReRF, we design a special FVV codec that achieves three orders of magnitudes compression rate and provides a companion ReRF player to support online streaming of long-duration FVVs of dynamic scenes. Extensive experiments demonstrate the effectiveness of ReRF for compactly representing dynamic radiance fields, enabling an unprecedented free-viewpoint viewing experience in speed and quality.
翻译:神经辐射场(NeRFs)在建模和自由视点渲染静态物体方面的成功启发了许多针对动态场景的尝试。目前利用神经渲染促进自由视点视频(FVVs)的技术限于离线渲染,或仅能处理最小运动的短序列。在本文中,我们提出了一种新型技术,即残差辐射场(ReRF),作为高度紧凑的神经表示,以实现长时动态场景的实时FVV渲染。ReRF在时空特征空间中显式地建模相邻时间戳之间的残差信息,使用全局基于坐标的微型MLP作为特征解码器。具体而言,ReRF使用紧凑的运动网格以及残差特征网格来利用帧间特征相似性,无需牺牲质量即可处理大运动。我们进一步提出了一种顺序训练方案,以维护运动/残差网格的平滑性和稀疏性。基于ReRF,我们设计了一个特殊的FVV编解码器,实现了三个数量级的压缩率,并提供了一个ReRF播放器,支持动态场景长时FVV的在线流式处理。广泛的实验证明了ReRF在紧凑表示动态辐射场方面的有效性,使速度和质量的自由视角查看体验达到前所未有的水平。