In many recent works, multi-layer perceptions (MLPs) have been shown to be suitable for modeling complex spatially-varying functions including images and 3D scenes. Although the MLPs are able to represent complex scenes with unprecedented quality and memory footprint, this expressive power of the MLPs, however, comes at the cost of long training and inference times. On the other hand, bilinear/trilinear interpolation on regular grid based representations can give fast training and inference times, but cannot match the quality of MLPs without requiring significant additional memory. Hence, in this work, we investigate what is the smallest change to grid-based representations that allows for retaining the high fidelity result of MLPs while enabling fast reconstruction and rendering times. We introduce a surprisingly simple change that achieves this task -- simply allowing a fixed non-linearity (ReLU) on interpolated grid values. When combined with coarse to-fine optimization, we show that such an approach becomes competitive with the state-of-the-art. We report results on radiance fields, and occupancy fields, and compare against multiple existing alternatives. Code and data for the paper are available at https://geometry.cs.ucl.ac.uk/projects/2022/relu_fields.
翻译:在许多近期的工程中,多层认知(MLPs)被证明适合于模拟复杂的空间变化功能,包括图像和3D场景。尽管MLPs能够代表具有前所未有的质量和记忆足迹的复杂场景,但MLPs的这种显性力量是以长期培训和推断时间的代价而来的。另一方面,常规电网代表制的双线/三线间插划可以提供快速的培训和推断时间,但无法在不需要大量额外记忆的情况下与MLPs的质量相匹配。因此,在这项工作中,我们调查对基于电网的表述进行最小的改变,使MLPs能够保持高度忠诚的结果,同时能够快速重建并创造时间。我们引入了一个令人惊讶的简单变化,即仅允许固定的非线性(ReLU)在内部电网值上进行。当与粗度到最优化相结合时,我们表明这种方式与状态-艺术具有竞争力,因此我们报告在直线字段和占用域的成果,并比较现有的纸质标准/模型。