The modern strategy for training deep neural networks for classification tasks includes optimizing the network's weights even after the training error vanishes to further push the training loss toward zero. Recently, a phenomenon termed "neural collapse" (NC) has been empirically observed in this training procedure. Specifically, it has been shown that the learned features (the output of the penultimate layer) of within-class samples converge to their mean, and the means of different classes exhibit a certain tight frame structure, which is also aligned with the last layer's weights. Recent papers have shown that minimizers with this structure emerge when optimizing a simplified "unconstrained features model" (UFM) with a regularized cross-entropy loss. In this paper, we further analyze and extend the UFM. First, we study the UFM for the regularized MSE loss, and show that the minimizers' features can have a more delicate structure than in the cross-entropy case. This affects also the structure of the weights. Then, we extend the UFM by adding another layer of weights as well as ReLU nonlinearity to the model and generalize our previous results. Finally, we empirically demonstrate the usefulness of our nonlinear extended UFM in modeling the NC phenomenon that occurs with practical networks.
翻译:为分类任务培训深神经网络的现代战略包括优化网络的重量,即使培训错误消失后网络的重量,以进一步将培训损失推向零。最近,在这一培训过程中,从经验中观察到了一个名为“神经崩溃”的现象。具体地说,已经表明,本类样本中学习到的特征(倒数第二层的产出)与其平均值一致,而不同类别的方法也表现出某种紧凑的框架结构,这也与最后一层的重量相一致。最近的文件表明,在优化简化的“不受限制的特征模型”(UFM)时,这一结构的最小化作用会显现出来,而简化的“不受限制的特征模型”(UFMM)则导致常规化的跨热带损失。在本文中,我们进一步分析并扩展了UFM。首先,我们为常规化的 MSE损失而研究UFM, 并表明,最小化器的特征结构可能比跨层的重量结构更为微妙。然后,我们通过增加另一层的重量来扩大UFMM,同时将 ReLU的非直线性化为模型,我们以往的实用性能模型最终展示了我们的UFM 。