The main difficulty that arises in the analysis of most machine learning algorithms is to handle, analytically and numerically, a large number of interacting random variables. In this Ph.D manuscript, we revisit an approach based on the tools of statistical physics of disordered systems. Developed through a rich literature, they have been precisely designed to infer the macroscopic behavior of a large number of particles from their microscopic interactions. At the heart of this work, we strongly capitalize on the deep connection between the replica method and message passing algorithms in order to shed light on the phase diagrams of various theoretical models, with an emphasis on the potential differences between statistical and algorithmic thresholds. We essentially focus on synthetic tasks and data generated in the teacher-student paradigm. In particular, we apply these mean-field methods to the Bayes-optimal analysis of committee machines, to the worst-case analysis of Rademacher generalization bounds for perceptrons, and to empirical risk minimization in the context of generalized linear models. Finally, we develop a framework to analyze estimation models with structured prior informations, produced for instance by deep neural networks based generative models with random weights.
翻译:分析大多数机器学习算法过程中出现的主要困难在于从分析和数字上处理大量互动随机变量。 在这份博士手稿中,我们重新审视了一种基于无序系统统计物理工具的方法。通过丰富的文献开发,它们精确地设计来推断微小相互作用中大量微粒的宏观行为。在这项工作的核心,我们大力利用复制方法与信息传递算法之间的深层联系,以便了解各种理论模型的阶段图,重点是统计和算法阈值之间的潜在差异。我们基本上侧重于教师-学生范式中产生的合成任务和数据。特别是,我们将这些平均领域方法应用于对委员会机器的巴耶斯最佳分析,用于对Rademacher的透视带进行最坏情况分析,并用于在一般线性模型中将经验风险降到最低。最后,我们开发了一个框架,以结构化的先前信息来分析模型,例如由基于随机重量的深层神经网络基因分析模型生成的模型。