PAC-Bayes learning is an established framework to assess the generalisation ability of learning algorithm during the training phase. However, it remains challenging to know whether PAC-Bayes is useful to understand, before training, why the output of well-known algorithms generalise well. We positively answer this question by expanding the \emph{Wasserstein PAC-Bayes} framework, briefly introduced in \cite{amit2022ipm}. We provide new generalisation bounds exploiting geometric assumptions on the loss function. Using our framework, we prove, before any training, that the output of an algorithm from \citet{lambert2022variational} has a strong asymptotic generalisation ability. More precisely, we show that it is possible to incorporate optimisation results within a generalisation framework, building a bridge between PAC-Bayes and optimisation algorithms.
翻译:PAC-Bayes学习是在训练阶段评估学习算法泛化能力的一种成熟框架。然而,在训练之前了解为什么著名算法的输出能够广泛地泛化仍然具有挑战性。我们通过扩展在\cite{amit2022ipm}中简要介绍的\emph{Wasserstein PAC-Bayes}框架来肯定回答了这个问题。我们提供了利用损失函数的几何假设的新的泛化界。利用我们的框架,我们在任何训练之前证明了\citet{lambert2022variational}中算法的输出具有强大的渐近泛化能力。更具体地说,我们展示了如何将优化结果纳入泛化框架中,构建PAC-Bayes和优化算法之间的桥梁。