With increasing deployment of machine learning systems in various real-world tasks, there is a greater need for accurate quantification of predictive uncertainty. While the common goal in uncertainty quantification (UQ) in machine learning is to approximate the true distribution of the target data, many works in UQ tend to be disjoint in the evaluation metrics utilized, and disparate implementations for each metric lead to numerical results that are not directly comparable across different works. To address this, we introduce Uncertainty Toolbox, an open-source python library that helps to assess, visualize, and improve UQ. Uncertainty Toolbox additionally provides pedagogical resources, such as a glossary of key terms and an organized collection of key paper references. We hope that this toolbox is useful for accelerating and uniting research efforts in uncertainty in machine learning.
翻译:随着在各种现实世界任务中越来越多地部署机器学习系统,更加需要准确量化预测性不确定性。虽然在机器学习中不确定性量化(UQ)的共同目标是接近目标数据的真正分布,但在UQ的许多工作往往与所使用的评价指标脱节,而且对每一公吨铅的不同实施导致不同工程之间无法直接比较的数字结果。为了解决这个问题,我们引入了不确定性工具箱,这是一个开放源码的皮松图书馆,有助于评估、可视化和改进UQ。不确定性工具箱提供额外的教学资源,例如关键术语词汇表和有组织地收集关键文件参考资料。我们希望这一工具箱有助于加速和集中研究机器学习中的不确定性。