The failure of deep neural networks to generalize to out-of-distribution data is a well-known problem and raises concerns about the deployment of trained networks in safety-critical domains such as healthcare, finance and autonomous vehicles. We study a particular kind of distribution shift $\unicode{x2013}$ shortcuts or spurious correlations in the training data. Shortcut learning is often only exposed when models are evaluated on real-world data that does not contain the same spurious correlations, posing a serious dilemma for AI practitioners to properly assess the effectiveness of a trained model for real-world applications. In this work, we propose to use the mutual information (MI) between the learned representation and the input as a metric to find where in training, the network latches onto shortcuts. Experiments demonstrate that MI can be used as a domain-agnostic metric for monitoring shortcut learning.
翻译:深层神经网络未能将数据推广到传播之外,这是一个众所周知的问题,令人对在保健、金融和自主车辆等安全关键领域部署经过培训的网络感到关切。我们研究了一种特殊的分发转移 $\ uncode{x2013}$ 捷径或培训数据中虚假的关联。当模型在现实世界数据上评价模型时,没有包含同样的虚假关联时,快速学习通常才会暴露出来,这对大赦国际从业人员正确评估经过培训的真实世界应用模型的有效性造成了严重的困境。在这项工作中,我们提议使用所学代表方与投入方之间的相互信息作为衡量标准,以查找培训中哪些地方的网络拉特切到捷径。实验表明MI可以用作监测捷径学习的域名衡量标准。