The goal of coreset selection in supervised learning is to produce a weighted subset of data, so that training only on the subset achieves similar performance as training on the entire dataset. Existing methods achieved promising results in resource-constrained scenarios such as continual learning and streaming. However, most of the existing algorithms are limited to traditional machine learning models. A few algorithms that can handle large models adopt greedy search approaches due to the difficulty in solving the discrete subset selection problem, which is computationally costly when coreset becomes larger and often produces suboptimal results. In this work, for the first time we propose a continuous probabilistic bilevel formulation of coreset selection by learning a probablistic weight for each training sample. The overall objective is posed as a bilevel optimization problem, where 1) the inner loop samples coresets and train the model to convergence and 2) the outer loop updates the sample probability progressively according to the model's performance. Importantly, we develop an efficient solver to the bilevel optimization problem via unbiased policy gradient without trouble of implicit differentiation. We provide the convergence property of our training procedure and demonstrate the superiority of our algorithm against various coreset selection methods in various tasks, especially in more challenging label-noise and class-imbalance scenarios.
翻译:在监督的学习中,核心设置选择的目标是产生一组加权数据,以便仅对子组进行的培训能够取得与整个数据集培训相似的性能。现有方法在持续学习和流学等资源受限制的情景中取得了有希望的成果。然而,大多数现有算法局限于传统的机器学习模式。由于难以解决离散子组选择问题,处理大型模型的少数算法可以采用贪婪的搜索方法。当核心数据集变大并经常产生不理想的结果时,这种算法成本很高。在这项工作中,我们首次通过学习每个培训样本的概率重量来提出核心设置的持续概率性双级配置。总体目标是双级优化问题,其中1) 内部循环核心设置样本,并训练模型的趋同,2) 外环根据模型的性能,逐步更新样本的概率。重要的是,我们开发一个高效的解决方案,通过公平的政策梯度来解决双级优化问题,而不产生隐含差异。我们提供了培训程序的趋同性,并展示了我们的算法优于不同任务中具有挑战性能的等级标签。