This work studies the design of neural networks that can process the weights or gradients of other neural networks, which we refer to as neural functional networks (NFNs). Despite a wide range of potential applications, including learned optimization, processing implicit neural representations, network editing, and policy evaluation, there are few unifying principles for designing effective architectures that process the weights of other networks. We approach the design of neural functionals through the lens of symmetry, in particular by focusing on the permutation symmetries that arise in the weights of deep feedforward networks because hidden layer neurons have no inherent order. We introduce a framework for building permutation equivariant neural functionals, whose architectures encode these symmetries as an inductive bias. The key building blocks of this framework are NF-Layers (neural functional layers) that we constrain to be permutation equivariant through an appropriate parameter sharing scheme. In our experiments, we find that permutation equivariant neural functionals are effective on a diverse set of tasks that require processing the weights of MLPs and CNNs, such as predicting classifier generalization, producing "winning ticket" sparsity masks for initializations, and editing the weights of implicit neural representations (INRs). In addition, we provide code for our models and experiments at https://github.com/AllanYangZhou/nfn.
翻译:这项工作研究能够处理其他神经网络的权重或梯度的神经网络的设计,我们称之为神经功能网络。尽管存在广泛的潜在应用,包括学习优化、处理隐含神经表示、网络编辑和政策评价,但设计有效结构以处理其他网络权重的一致原则却很少。我们通过对称透镜来看待神经功能的设计,特别是通过适当的参数共享计划,侧重于深层饲料网络重量中产生的变异性对称,因为隐藏的层神经没有内在的顺序。我们引入了一个建设变异等等神经功能的框架,其结构将这些对等性作为一种感性偏差进行编码。这个框架的关键构件是NF-Layers(神经功能层),我们通过适当的参数共享计划来限制变异性。在我们的实验中,我们发现变异性神经功能对多种任务组合是有效的,这一系列任务都要求对内置的变异性模型进行精度分析,对内置的精度进行精度预测,对内存的内嵌性分析,例如,对内存的内置的内置分析,对内置的模型进行精度分析,对内置的精度分析,例如,对内存的内存的内存和内置的内置的内置的模拟,对等。</s>