Influence functions efficiently estimate the effect of removing a single training data point on a model's learned parameters. While influence estimates align well with leave-one-out retraining for linear models, recent works have shown this alignment is often poor in neural networks. In this work, we investigate the specific factors that cause this discrepancy by decomposing it into five separate terms. We study the contributions of each term on a variety of architectures and datasets and how they vary with factors such as network width and training time. While practical influence function estimates may be a poor match to leave-one-out retraining for nonlinear networks, we show they are often a good approximation to a different object we term the proximal Bregman response function (PBRF). Since the PBRF can still be used to answer many of the questions motivating influence functions, such as identifying influential or mislabeled examples, our results suggest that current algorithms for influence function estimation give more informative results than previous error analyses would suggest.
翻译:影响函数有效估计了删除单一培训数据点对模型所学参数的影响。虽然影响估计与线性模型的一出休假再培训非常吻合,但最近的工程显示,神经网络中的这种配合往往很差。在这项工作中,我们调查造成这种差异的具体因素,将其分为五个不同的术语。我们研究每个术语对各种结构和数据集的贡献,以及它们与网络宽度和培训时间等因素的差异。虽然实际影响函数估计可能与非线性网络的一出休假再培训不匹配,但我们表明,它们往往与我们称之为Proximal Bregman反应功能(PBRF)的不同对象相近。由于PBRF仍可用于回答许多激发影响功能的问题,例如确定有影响力或标签错误的例子,我们的结果表明,目前用于影响函数估计的算法比以往的错误分析所显示的要具有更多信息。