Initialising the synaptic weights of artificial neural networks (ANNs) with orthogonal matrices is known to alleviate vanishing and exploding gradient problems. A major objection against such initialisation schemes is that they are deemed biologically implausible as they mandate factorization techniques that are difficult to attribute to a neurobiological process. This paper presents two initialisation schemes that allow a network to naturally evolve its weights to form orthogonal matrices, provides theoretical analysis that pre-training orthogonalisation always converges, and empirically confirms that the proposed schemes outperform randomly initialised recurrent and feedforward networks.
翻译:以正向矩阵初始化人工神经网络(ANNs)的合成权重,已知可以缓解消失和爆炸梯度问题,对此类初始化计划的主要反对意见是,它们被认为在生物学上不可信,因为它们要求采用难以归属于神经生物过程的乘数化技术。本文介绍了两个初始化计划,使网络能够自然地将其重量演变成正向矩阵,提供了理论分析,表明培训前或调整总是会趋同,并用经验证实,拟议的计划超越了随机创建的经常性和进料式网络。