In this work, we propose ModelPred, a framework that helps to understand the impact of changes in training data on a trained model. This is critical for building trust in various stages of a machine learning pipeline: from cleaning poor-quality samples and tracking important ones to be collected during data preparation, to calibrating uncertainty of model prediction, to interpreting why certain behaviors of a model emerge during deployment. Specifically, ModelPred learns a parameterized function that takes a dataset $S$ as the input and predicts the model obtained by training on $S$. Our work differs from the recent work of Datamodels [1] as we aim for predicting the trained model parameters directly instead of the trained model behaviors. We demonstrate that a neural network-based set function class is capable of learning the complex relationships between the training data and model parameters. We introduce novel global and local regularization techniques to prevent overfitting and we rigorously characterize the expressive power of neural networks (NN) in approximating the end-to-end training process. Through extensive empirical investigations, we show that ModelPred enables a variety of applications that boost the interpretability and accountability of machine learning (ML), such as data valuation, data selection, memorization quantification, and model calibration.
翻译:在这项工作中,我们提出模型预测,这是一个有助于理解培训数据变化对经过培训的模型的影响的框架。这对于在机器学习管道的各个阶段建立信任至关重要:从清理质量差的样本和跟踪在数据准备期间收集的重要样本,到校准模型预测的不确定性,以及解释模型部署期间出现某些行为的原因。具体地说,模型学习一个参数化功能,将数据集的美元作为投入,并预测通过培训获得的模型。我们的工作不同于数据模型[1]的近期工作[1],因为我们的目标是直接预测经过培训的模型参数,而不是经过培训的模型行为。我们证明,基于神经网络的功能班能够学习培训数据与模型参数之间的复杂关系。我们采用了新的全球和地方规范化技术,以防止过度匹配,我们严格地描述神经网络(NNN)在适应端到端培训过程中的表达力。通过广泛的实证调查,我们表明模型化能够使各种应用能够促进机器学习的可解释性和问责性、数据评估(ML)等数据,例如数据评估。