With a recent observation of the "Neural Collapse (NC)" phenomena by Papyan et al., various efforts have been made to model it and analyse the implications. Neural collapse describes that in deep classifier networks, the class features of the final hidden layer associated with training data tend to collapse to the respective class feature means. Thus, simplifying the behaviour of the last layer classifier to that of a nearest-class center decision rule. In this work, we analyse the principles which aid in modelling such a phenomena from the ground up and show how they can build a common understanding of the recently proposed models that try to explain NC. We hope that our analysis presents a multifaceted perspective on modelling NC and aids in forming connections with the generalization capabilities of neural networks. Finally, we conclude by discussing the avenues for further research and propose potential research problems.
翻译:Papyan等人最近观察了“Neal Craw(NC)”现象,因此作出了各种努力来模拟和分析其影响。神经崩溃说明,在深层分类网络中,与培训数据相关的最后隐藏层的等级特征往往会崩溃到相应的类别特征。因此,将最后一层分类者的行为简化为近层中心决策规则的行为。在这项工作中,我们分析有助于从头开始模拟这种现象的原则,并表明它们如何能够对最近提出的、试图解释NC的模式形成共同的理解。我们希望我们的分析能够从多方面的角度介绍NC的模型,协助建立与神经网络一般化能力的联系。最后,我们通过讨论进一步研究的途径和提出潜在的研究问题来结束我们的工作。