This paper presents an approach to addressing the issue of over-parametrization in deep neural networks, more specifically by avoiding the ``sparse double descent'' phenomenon. The authors propose a learning framework that allows avoidance of this phenomenon and improves generalization, an entropy measure to provide more insights on its insurgence, and provide a comprehensive quantitative analysis of various factors such as re-initialization methods, model width and depth, and dataset noise. The proposed approach is supported by experimental results achieved using typical adversarial learning setups. The source code to reproduce the experiments is provided in the supplementary materials and will be publicly released upon acceptance of the paper.
翻译:本文件提出了解决深层神经网络中过度平衡问题的办法,更具体地说,是避免“偏差的双血双血”现象,作者提出一个学习框架,以便避免这种现象,改进一般化,这是一种能使人们更深入了解其暴动的通缩措施,对重新启用方法、模型宽度和深度以及数据集噪音等各种因素进行全面的定量分析,提议的办法得到利用典型的对抗性学习组合取得的实验结果的支持,复制试验的源代码载于补充材料中,在文件被接受后将公开发布。</s>