Computational cost to train state-of-the-art deep models in many learning problems is rapidly increasing due to more sophisticated models and larger datasets. A recent promising direction to reduce training time is dataset condensation that aims to replace the original large training set with a significantly smaller learned synthetic set while preserving its information. While training deep models on the small set of condensed images can be extremely fast, their synthesis remains computationally expensive due to the complex bi-level optimization and second-order derivative computation. In this work, we propose a simple yet effective dataset condensation technique that requires significantly lower training cost with comparable performance by matching feature distributions of the synthetic and original training images in sampled embedding spaces. Thanks to its efficiency, we apply our method to more realistic and larger datasets with sophisticated neural architectures and achieve a significant performance boost while using larger synthetic training set. We also show various practical benefits of our method in continual learning and neural architecture search.
翻译:在许多学习问题中,培训最先进的深层模型的计算成本由于更先进的模型和更大的数据集而迅速增加。最近一个减少培训时间的有希望的方向是数据集凝结,目的是用一个小得多的学习合成数据集来取代原有的大型培训,同时保存信息。虽然对小型精密图像集的深层模型的培训速度极快,但由于复杂的双级优化和二级衍生计算,它们的合成仍然计算费用昂贵。在这项工作中,我们提出一种简单而有效的数据集凝聚技术,要求通过将抽样嵌入空间的合成和原始培训图像的特征分布相匹配,大大降低培训成本和可比较性能,从而大大降低培训成本。由于它的效率,我们运用了我们的方法,用复杂的神经结构来更现实、更大型的数据集,并在使用更大的合成培训集中实现显著的性能提升。我们还展示了我们方法在持续学习和神经结构搜索方面的各种实际效益。