Neural networks are known to use spurious correlations such as background information for classification. While prior work has looked at spurious correlations that are widespread in the training data, in this work, we investigate how sensitive neural networks are to rare spurious correlations, which may be harder to detect and correct, and may lead to privacy leaks. We introduce spurious patterns correlated with a fixed class to a few training examples and find that it takes only a handful of such examples for the network to learn the correlation. Furthermore, these rare spurious correlations also impact accuracy and privacy. We empirically and theoretically analyze different factors involved in rare spurious correlations and propose mitigation methods accordingly. Specifically, we observe that $\ell_2$ regularization and adding Gaussian noise to inputs can reduce the undesirable effects. Code available at https://github.com/yangarbiter/rare-spurious-correlation.
翻译:已知神经网络使用虚假的关联,如背景信息,进行分类。虽然先前的工作研究了培训数据中普遍存在的虚假关联,但在这项工作中,我们调查敏感神经网络如何成为稀有的虚假关联,这可能会更难发现和纠正,并可能导致隐私泄漏。我们引入了与固定类别相关的假模式,与少数培训实例相关联,发现网络学习这些关联只需要少数几个此类实例。此外,这些罕见的虚假关联也影响到准确性和隐私。我们从经验上和理论上分析了稀有假关联中涉及的不同因素,并据此提出缓解方法。具体地说,我们观察到,$@ell_2$的正规化和增加高斯噪音用于投入可以减少不良效应。代码见https://github.com/yangarbiter/rare-spurous-corlation-crelation。