Calibration of neural networks is a topical problem that is becoming more and more important as neural networks increasingly underpin real-world applications. The problem is especially noticeable when using modern neural networks, for which there is a significant difference between the confidence of the model and the probability of correct prediction. Various strategies have been proposed to improve calibration, yet accurate calibration remains challenging. We propose a novel framework with two contributions: introducing a differentiable surrogate for expected calibration error (DECE) that allows calibration quality to be directly optimised, and a meta-learning framework that uses DECE to optimise for validation set calibration with respect to model hyper-parameters. The results show that we achieve competitive performance with state-of-the-art calibration approaches. Our framework opens up a new avenue and toolset for tackling calibration, which we believe will inspire further work in this important challenge.
翻译:随着神经网络日益成为现实世界应用的基础,神经网络的校准是一个日新月异的热点问题,随着神经网络日益成为现实世界应用的基础,这一问题在使用现代神经网络时尤其明显,因为模型的信心与正确预测的概率之间存在巨大差异。提出了各种战略来改进校准,但精确校准仍然具有挑战性。我们提出了一个具有两种贡献的新框架:引入一个允许直接优化校准质量的可区别替代校准误差(DECE),以及一个利用DECE来优化校准模型超参数校准的元学习框架。结果显示,我们以最先进的校准方法实现了竞争性业绩。我们的框架为处理校准问题开辟了新的途径和工具,我们认为这将激励在这一重要挑战中开展进一步的工作。