The double descent curve is one of the most intriguing properties of deep neural networks. It contrasts the classical bias-variance curve with the behavior of modern neural networks, occurring where the number of samples nears the number of parameters. In this work, we explore the connection between the double descent phenomena and the number of samples in the deep neural network setting. In particular, we propose a construction which augments the existing dataset by artificially increasing the number of samples. This construction empirically mitigates the double descent curve in this setting. We reproduce existing work on deep double descent, and observe a smooth descent into the overparameterized region for our construction. This occurs both with respect to the model size, and with respect to the number epochs.
翻译:双下降曲线是深神经网络最令人感兴趣的特性之一,它与传统偏向偏差曲线与现代神经网络的行为形成对比,现代神经网络的样本数量接近参数数量。在这项工作中,我们探讨了双下降现象与深神经网络设置中的样本数量之间的联系。特别是,我们建议进行一项工程,通过人工增加样本数量来增加现有数据集。这一工程从经验上减轻了这一环境中的双下降曲线。我们复制了现有的深双下降工作,并观察到我们建筑工程在超分界线区域中平稳下降。这既涉及模型大小,也涉及数字时代。