Deep neural networks have demonstrated remarkable performance in supervised learning tasks but require large amounts of labeled data. Self-supervised learning offers an alternative paradigm, enabling the model to learn from data without explicit labels. Information theory has been instrumental in understanding and optimizing deep neural networks. Specifically, the information bottleneck principle has been applied to optimize the trade-off between compression and relevant information preservation in supervised settings. However, the optimal information objective in self-supervised learning remains unclear. In this paper, we review various approaches to self-supervised learning from an information-theoretic standpoint and present a unified framework that formalizes the \textit{self-supervised information-theoretic learning problem}. We integrate existing research into a coherent framework, examine recent self-supervised methods, and identify research opportunities and challenges. Moreover, we discuss empirical measurement of information-theoretic quantities and their estimators. This paper offers a comprehensive review of the intersection between information theory, self-supervised learning, and deep neural networks.
翻译:深度神经网络在监督学习任务中展示了出色的性能,但需要大量的标记数据。自监督学习提供了一种替代范式,使模型能够在没有明确标记的情况下从数据中学习。信息论在理解和优化深度神经网络方面发挥了重要作用。具体来说,信息瓶颈原理已被用于优化监督设置中压缩和相关信息保留之间的权衡。然而,在自监督学习中的最优信息目标仍不清楚。在本文中,我们从信息论的角度回顾了各种自监督学习方法,并提出了一个统一的框架,将“自监督信息论学习问题”形式化。我们将现有的研究整合成一个一致的框架,审查最近的自监督方法,并确定研究机会和挑战。此外,我们还讨论了信息论量和其估计器的经验度量。本文全面综述了信息论、自监督学习和深度神经网络之间的交叉领域。