Neural radiance fields (NeRF) have demonstrated the potential of coordinate-based neural representation (neural fields or implicit neural representation) in neural rendering. However, using a multi-layer perceptron (MLP) to represent a 3D scene or object requires enormous computational resources and time. There have been recent studies on how to reduce these computational inefficiencies by using additional data structures, such as grids or trees. Despite the promising performance, the explicit data structure necessitates a substantial amount of memory. In this work, we present a method to reduce the size without compromising the advantages of having additional data structures. In detail, we propose using the wavelet transform on grid-based neural fields. Grid-based neural fields are for fast convergence, and the wavelet transform, whose efficiency has been demonstrated in high-performance standard codecs, is to improve the parameter efficiency of grids. Furthermore, in order to achieve a higher sparsity of grid coefficients while maintaining reconstruction quality, we present a novel trainable masking approach. Experimental results demonstrate that non-spatial grid coefficients, such as wavelet coefficients, are capable of attaining a higher level of sparsity than spatial grid coefficients, resulting in a more compact representation. With our proposed mask and compression pipeline, we achieved state-of-the-art performance within a memory budget of 2 MB. Our code is available at https://github.com/daniel03c1/masked_wavelet_nerf.
翻译:神经光亮场( NERF) 展示了以协调为基础的神经神经显示( 神经字段或隐含神经代表) 在神经转化中的潜力。 但是,使用多层天体( MLP) 代表三维场或物体需要巨大的计算资源和时间。 最近对如何通过使用诸如电网或树木等额外数据结构来减少这些计算效率低下的问题进行了研究。尽管表现良好,但明确的数据结构需要大量的记忆。 在这项工作中,我们提出了一个在不损害拥有额外数据结构的优势的情况下缩小空间空间的方法。 详细来说,我们提议在基于电网的神经场上使用波变换。 基于网格的神经场是快速趋同的,而其效率已在高性能标准代码中显示出来提高电网的参数效率。此外,为了在保持重建质量的同时实现更高的电网系数,我们提出了一个新的可训练的掩码。 实验结果表明,非空间网格的系数,例如波变压的电动神经基数,能够达到我们预算内一个更高的水平。