It is common practice in deep learning to represent a measurement of the world on a discrete grid, e.g. a 2D grid of pixels. However, the underlying signal represented by these measurements is often continuous, e.g. the scene depicted in an image. A powerful continuous alternative is then to represent these measurements using an implicit neural representation, a neural function trained to output the appropriate measurement value for any input spatial location. In this paper, we take this idea to its next level: what would it take to perform deep learning on these functions instead, treating them as data? In this context we refer to the data as functa, and propose a framework for deep learning on functa. This view presents a number of challenges around efficient conversion from data to functa, compact representation of functa, and effectively solving downstream tasks on functa. We outline a recipe to overcome these challenges and apply it to a wide range of data modalities including images, 3D shapes, neural radiance fields (NeRF) and data on manifolds. We demonstrate that this approach has various compelling properties across data modalities, in particular on the canonical tasks of generative modeling, data imputation, novel view synthesis and classification.
翻译:深层学习通常的做法是在离散网格上代表世界的测量,例如2D象素网格。然而,这些测量所代表的基本信号往往是连续的,例如图像中描绘的场景。然后,一个强有力的连续的替代方法是,使用隐含的神经表象来代表这些测量,这是一种神经功能,经过培训,可以输出任何输入空间位置的适当测量值。在本文中,我们把这个想法带到下一个层次:要对这些功能进行深度学习,而不是将它们作为数据来对待,需要做什么?在这方面,我们把数据称为functa,并提出一个对functa进行深层学习的框架。这个观点提出了围绕从数据到functa、Functa的缩缩影和有效解决Functa上下游任务的有效转换的若干挑战。我们概述了克服这些挑战的配方,并将其应用于广泛的数据模式,包括图像、3D形状、神经光场(NERF)和多元数据数据。我们证明,这一方法在数据模式上具有各种令人信服的特性,特别是在基因化模型、数据合成和新分类方面。