To facilitate a wide-spread acceptance of AI systems guiding decision making in real-world applications, trustworthiness of deployed models is key. That is, it is crucial for predictive models to be uncertainty-aware and yield well-calibrated (and thus trustworthy) predictions for both in-domain samples as well as under domain shift. Recent efforts to account for predictive uncertainty include post-processing steps for trained neural networks, Bayesian neural networks as well as alternative non-Bayesian approaches such as ensemble approaches and evidential deep learning. Here, we propose an efficient yet general modelling approach for obtaining well-calibrated, trustworthy probabilities for samples obtained after a domain shift. We introduce a new training strategy combining an entropy-encouraging loss term with an adversarial calibration loss term and demonstrate that this results in well-calibrated and technically trustworthy predictions for a wide range of domain drifts. We comprehensively evaluate previously proposed approaches on different data modalities, a large range of data sets including sequence data, network architectures and perturbation strategies. We observe that our modelling approach substantially outperforms existing state-of-the-art approaches, yielding well-calibrated predictions under domain drift.
翻译:为了促进广泛接受指导现实应用决策的AI系统,已部署模型的可信度是关键所在。也就是说,预测模型必须是具有不确定性的,并且对内部样品和域变轨进行精确(因而可信)的预测。最近对预测不确定性的核算努力包括受过训练的神经网络的后处理步骤、贝耶斯神经网络以及替代的非巴伊西亚方法,如共同方法和证据深度学习。在这里,我们提出一个高效而又普遍的建模方法,以便为在域变换后取得的样品获得合理校准、可信赖的概率。我们引入了一个新的培训战略,将诱导损失术语与对抗校准损失术语结合起来,并表明这导致对广泛的域流流流进行充分校准和技术上可信赖的预测。我们全面评价了以前提出的不同数据模式、包括序列数据、网络架构和扰动战略在内的大量数据集的方法。我们观察到,我们的建模方法大大超越了现有区域流动预测。