The promising performances of CNNs often overshadow the need to examine whether they are doing in the way we are actually interested. We show through experiments that even over-parameterized models would still solve a dataset by recklessly leveraging spurious correlations, or so-called 'shortcuts'. To combat with this unintended propensity, we borrow the idea of printer test page and propose a novel approach called White Paper Assistance. Our proposed method involves the white paper to detect the extent to which the model has preference for certain characterized patterns and alleviates it by forcing the model to make a random guess on the white paper. We show the consistent accuracy improvements that are manifest in various architectures, datasets and combinations with other techniques. Experiments have also demonstrated the versatility of our approach on fine-grained recognition, imbalanced classification and robustness to corruptions.
翻译:CNN有希望的表现往往掩盖了审查它们是否以我们实际感兴趣的方式行事的必要性。我们通过实验显示,即使是超分度模型也会通过鲁莽地利用虚假的关联或所谓的“短切”来解决数据集问题。为了与这种无意的倾向作斗争,我们借用了打印机测试页的想法,并提出了一个叫作“白皮书援助”的新颖方法。我们建议的方法涉及白皮书,以检测模型偏好某些特征模式的程度,并通过迫使模型随机猜测白皮书来缓解这一模式。我们展示了各种结构、数据集和与其他技术的结合所显示的一贯准确性改进。实验还展示了我们在微小识别、不平衡分类和腐败稳健方面的做法的多功能性。