The acquisition of advanced models relies on large datasets in many fields, which makes storing datasets and training models expensive. As a solution, dataset distillation can synthesize a small dataset such that models trained on it achieve high performance on par with the original large dataset. The recently proposed dataset distillation method by matching network parameters has been proved effective for several datasets. However, a few parameters in the distillation process are difficult to match, which harms the distillation performance. Based on this observation, this paper proposes a new method to solve the problem using parameter pruning. The proposed method can synthesize more robust distilled datasets and improve the distillation performance by pruning difficult-to-match parameters in the distillation process. Experimental results on three datasets show that the proposed method outperformed other SOTA dataset distillation methods.
翻译:先进模型的获取取决于许多领域的大型数据集,这使得存储数据集和培训模型的费用昂贵。作为一个解决方案,数据集蒸馏可以合成一个小数据集,使经过培训的模型在与原始大型数据集相当的程度上能取得高性能。最近提出的通过匹配网络参数进行数据集蒸馏的方法已证明对若干数据集有效。然而,蒸馏过程中的一些参数难以匹配,从而损害蒸馏性能。根据这一观察,本文件提出一种新方法,用参数修剪来解决问题。拟议方法可以通过在蒸馏过程中修剪难到匹配的参数,合成更强的蒸馏数据集,改进蒸馏性能。三个数据集的实验结果表明,拟议方法比其他SOTA数据蒸馏方法要好。