People usually believe that network pruning not only reduces the computational cost of deep networks, but also prevents overfitting by decreasing model capacity. However, our work surprisingly discovers that network pruning sometimes even aggravates overfitting. We report an unexpected sparse double descent phenomenon that, as we increase model sparsity via network pruning, test performance first gets worse (due to overfitting), then gets better (due to relieved overfitting), and gets worse at last (due to forgetting useful information). While recent studies focused on the deep double descent with respect to model overparameterization, they failed to recognize that sparsity may also cause double descent. In this paper, we have three main contributions. First, we report the novel sparse double descent phenomenon through extensive experiments. Second, for this phenomenon, we propose a novel learning distance interpretation that the curve of $\ell_{2}$ learning distance of sparse models (from initialized parameters to final parameters) may correlate with the sparse double descent curve well and reflect generalization better than minima flatness. Third, in the context of sparse double descent, a winning ticket in the lottery ticket hypothesis surprisingly may not always win.
翻译:人们通常认为,网络运行不仅会降低深层网络的计算成本,而且会防止模型容量的下降,但是,我们的工作令人惊讶地发现,网络运行有时甚至会加剧过度装配。我们报告出乎意料的稀疏双向下降现象,随着我们通过网络运行增加模型宽度,测试性能首先变得更差(由于过度装配),然后(由于过度装配)变得更好,最后(由于忘记了有用的信息)变得更差。虽然最近的研究侧重于模型超分化的深度双向下降,但他们没有认识到过度性也可能导致双向下降。在本文中,我们有三个主要贡献。首先,我们通过广泛的实验报告新颖的稀有双向下降现象。第二,我们提出一个新的远距离解释,即稀释型模型的曲线(从初始参数到最终参数)可能与稀释的双向下降曲线相关联,并且反映一般化比小幅平坦。第三,在稀疏的双向下降的背景下,在彩票的假设中赢票可能永远不会赢。