Neural networks have achieved impressive results on many technological and scientific tasks. Yet, their empirical successes have outpaced our fundamental understanding of their structure and function. Identifying mechanisms driving the successes of neural networks can provide principled approaches for improving neural network performance and developing simple and effective alternatives. In this work, we isolate a key mechanism driving feature learning in fully connected neural networks by connecting neural feature learning to a statistical estimator known as average gradient outer product. We subsequently leverage this mechanism to design \textit{Recursive Feature Machines} (RFMs), which are kernel machines that learn features. We show that RFMs (1) accurately capture features learned by deep fully connected neural networks, and (2) outperform a broad spectrum of models including neural networks on tabular data. Furthermore, we show how RFMs shed light on recently observed deep learning phenomena including grokking, lottery tickets, simplicity biases, and spurious features. We provide a Python implementation to make our method easily accessible [\url{https://github.com/aradha/recursive_feature_machines}].
翻译:在许多技术和科学任务上,神经网络取得了令人印象深刻的成果,然而,它们的实证成功超过了我们对自身结构和功能的基本理解。确定神经网络成功推动机制可以提供改善神经网络性能和开发简单有效的替代方法的原则性方法。在这项工作中,我们通过将神经特征学习与一个称为平均梯度外产品的统计估计器联系起来,将完全连通的神经网络学习的关键机制分离出来。我们随后利用这一机制设计了具有学习特点的内核机器。我们表明RFM(1) 准确地捕捉了完全连接的神经网络所学的特征,以及(2) 超越了广泛的模型范围,包括表格数据的神经网络。此外,我们展示了RFM如何揭示最近观察到的深层学习现象,包括石刻、彩票、简单偏差和令人毛骨悚然的特征。我们提供了一种Python实施方法,使我们的方法更容易获得[https://github.com/aradha/recursivisive_fegramas}。