Good models require good training data. For overparameterized deep models, the causal relationship between training data and model predictions is increasingly opaque and poorly understood. Influence analysis partially demystifies training's underlying interactions by quantifying the amount each training instance alters the final model. Measuring the training data's influence exactly can be provably hard in the worst case; this has led to the development and use of influence estimators, which only approximate the true influence. This paper provides the first comprehensive survey of training data influence analysis and estimation. We begin by formalizing the various, and in places orthogonal, definitions of training data influence. We then organize state-of-the-art influence analysis methods into a taxonomy; we describe each of these methods in detail and compare their underlying assumptions, asymptotic complexities, and overall strengths and weaknesses. Finally, we propose future research directions to make influence analysis more useful in practice as well as more theoretically and empirically sound. A curated, up-to-date list of resources related to influence analysis is available at https://github.com/ZaydH/influence_analysis_papers.
翻译:良好的模型需要良好的培训数据。对于过于精确的深层模型来说,培训数据与模型预测之间的因果关系越来越不透明,而且不易理解。影响分析通过量化每个培训实例的数量改变最后模型,部分地淡化了培训的基本相互作用。衡量培训数据的影响,在最坏的情况下,完全可以看得出来是困难的;这导致了影响估计器的开发和使用,而影响估计器只是接近真正的影响。本文件提供了对培训数据影响分析和估计的第一次全面调查。我们首先将培训数据影响的各种定义和在正方位的定义正式化。我们然后将最新的影响分析方法组织成一种分类法;我们详细描述每一种方法,比较其基本假设、症状复杂性以及总体优缺点。最后,我们提出今后的研究方向,使影响分析在实践中更有用,并在理论上和经验上更可靠。与影响分析有关的最新资源清单见https://github.com/ZaydH/implication_paper。