Neural radiance fields (NeRF) has achieved outstanding performance in modeling 3D objects and controlled scenes, usually under a single scale. In this work, we focus on multi-scale cases where large changes in imagery are observed at drastically different scales. This scenario vastly exists in real-world 3D environments, such as city scenes, with views ranging from satellite level that captures the overview of a city, to ground level imagery showing complex details of an architecture; and can also be commonly identified in landscape and delicate minecraft 3D models. The wide span of viewing positions within these scenes yields multi-scale renderings with very different levels of detail, which poses great challenges to neural radiance field and biases it towards compromised results. To address these issues, we introduce BungeeNeRF, a progressive neural radiance field that achieves level-of-detail rendering across drastically varied scales. Starting from fitting distant views with a shallow base block, as training progresses, new blocks are appended to accommodate the emerging details in the increasingly closer views. The strategy progressively activates high-frequency channels in NeRF's positional encoding inputs and successively unfolds more complex details as the training proceeds. We demonstrate the superiority of BungeeNeRF in modeling diverse multi-scale scenes with drastically varying views on multiple data sources (city models, synthetic, and drone captured data) and its support for high-quality rendering in different levels of detail.
翻译:在3D天体和受控场景的模型模型模型(NERF)中,神经光亮场(NERF)通常在单一尺度下取得了杰出的性能。在这项工作中,我们侧重于在非常不同尺度上观测到图像发生巨大变化的多尺度案例。这种情景在现实世界的3D环境中存在,例如城市景色,从卫星水平到地面图像,从反映城市的概况到显示一个建筑的复杂细节的浅层图像,都可以在地貌和微妙的3D模型中发现。这些场景的广度观察位置产生不同细节的多级图像,给神经光亮场带来巨大的挑战,并偏向损害的结果。为了解决这些问题,我们引入Bungee NERF,一个进步的神经光亮场,一个渐进的光亮度在高度上跨越了千差万变万变的尺度,从一个浅浅浅的基块的远处开始,随着培训的进展,新的区块被附着,以适应日益接近的视角中出现的新细节。这一战略在NRF的定位模型中逐步启动高频频道,对神经光谱域域域域域域域域域域域域域域域的输入,并连续地展示各种的高级数据数据,以显示。