We analyze the epistemic uncertainty (EU) of supervised learning in Bayesian inference by focusing on the excess risk. Existing analysis is limited to the Bayesian setting, which assumes a correct model and exact Bayesian posterior distribution. Thus we cannot apply the existing theory to modern Bayesian algorithms, such as variational inference. To address this, we present a novel EU analysis in the frequentist setting, where data is generated from an unknown distribution. We show a relation between the generalization ability and the widely used EU measurements, such as the variance and entropy of the predictive distribution. Then we show their convergence behaviors theoretically. Finally, we propose new variational inference that directly controls the prediction and EU evaluation performances based on the PAC-Bayesian theory. Numerical experiments show that our algorithm significantly improves the EU evaluation over the existing methods.
翻译:我们通过注重过度风险来分析贝叶斯推论中受监督的认知性不确定性(EU),现有分析仅限于贝叶斯环境,该环境假定了正确的模型和精确的贝叶斯后子分布。因此我们不能将现有理论应用于现代贝叶斯算法,例如变式推论。为了解决这个问题,我们在常客环境中提出了欧盟的新分析,数据来自未知分布。我们显示了一般化能力与广泛使用的欧盟测量方法(如预测分布的差异和酶)之间的关系。然后我们从理论上展示了它们的趋同行为。最后,我们提出了新的变式推论,直接控制基于PAC-巴耶斯理论的预测和欧盟评估业绩。数字实验表明,我们的算法极大地改进了欧盟对现有方法的评估。