Associative Memories like the famous Hopfield Networks are elegant models for describing fully recurrent neural networks whose fundamental job is to store and retrieve information. In the past few years they experienced a surge of interest due to novel theoretical results pertaining to their information storage capabilities, and their relationship with SOTA AI architectures, such as Transformers and Diffusion Models. These connections open up possibilities for interpreting the computation of traditional AI networks through the theoretical lens of Associative Memories. Additionally, novel Lagrangian formulations of these networks make it possible to design powerful distributed models that learn useful representations and inform the design of novel architectures. This tutorial provides an approachable introduction to Associative Memories, emphasizing the modern language and methods used in this area of research, with practical hands-on mathematical derivations and coding notebooks.
翻译:著名的Hopfield网络等关联记忆模型,是描述全循环神经网络的优雅模型,其基本功能是存储和检索信息。过去几年,由于关于其信息存储能力的新理论成果,及其与Transformers和扩散模型等最先进人工智能架构的关联,这类模型重新引起了广泛关注。这些关联为通过关联记忆的理论视角解释传统人工智能网络的计算开辟了可能性。此外,这些网络的新拉格朗日公式使得设计强大的分布式模型成为可能,这些模型能够学习有用的表征,并为新型架构的设计提供参考。本教程对关联记忆进行了通俗易懂的介绍,重点阐述了该研究领域使用的现代语言和方法,并提供了实用的数学推导和代码实践。