Deep neural networks are highly performant, but might base their decision on spurious or background features that co-occur with certain classes, which can hurt generalization. To mitigate this issue, the usage of 'model guidance' has gained popularity recently: for this, models are guided to be "right for the right reasons" by regularizing the models' explanations to highlight the right features. Experimental validation of these approaches has thus far however been limited to relatively simple and / or synthetic datasets. To gain a better understanding of which model-guiding approaches actually transfer to more challenging real-world datasets, in this work we conduct an in-depth evaluation across various loss functions, attribution methods, models, and 'guidance depths' on the PASCAL VOC 2007 and MS COCO 2014 datasets, and show that model guidance can sometimes even improve model performance. In this context, we further propose a novel energy loss, show its effectiveness in directing the model to focus on object features. We also show that these gains can be achieved even with a small fraction (e.g. 1%) of bounding box annotations, highlighting the cost effectiveness of this approach. Lastly, we show that this approach can also improve generalization under distribution shifts. Code will be made available.
翻译:深度神经网络具有高性能,但可能会基于与某些类别同时出现的虚假或背景特征来做出决策,这可能会损害泛化能力。为了缓解这个问题,最近流行起来的“模型指导”方法:通过规范化模型的解释来突出正确的特征,指导模型“出于正确的原因而正确”。然而,这些方法的实验验证迄今为止仅限于相对简单和/或合成数据集。为了更好地了解哪些模型指导方法实际上适用于更具挑战性的真实世界数据集,我们在PASCAL VOC 2007和MS COCO 2014数据集上进行了全面的评估。在不同的损失函数、归因方法、模型和“指导深度”之间进行对比,表明模型指导有时甚至可以提高模型性能。在这种情况下,我们提出了一种新的能量损失函数,并展示了它在引导模型专注于对象特征方面的有效性。我们还展示了即使在少量笔记框注释(例如1%)的情况下,也可以实现这些收益,突出了该方法的成本效益。最后,我们表明,该方法也可以提高分布转换下的泛化能力。代码将可用。