This paper proposes the Mesh Neural Network (MNN), a novel architecture which allows neurons to be connected in any topology, to efficiently route information. In MNNs, information is propagated between neurons throughout a state transition function. State and error gradients are then directly computed from state updates without backward computation. The MNN architecture and the error propagation schema is formalized and derived in tensor algebra. The proposed computational model can fully supply a gradient descent process, and is potentially suitable for very large scale sparse NNs, due to its expressivity and training efficiency, with respect to NNs based on back-propagation and computational graphs.
翻译:本文建议采用Mesh神经网络(MNN),这是一个允许神经元在任何地形上连接的新型结构,可以有效地传递信息。在MNNS中,信息在神经元之间在整个国家过渡功能中传播。然后直接从国家更新中计算国家和错误梯度,不进行后向计算。MNN结构和错误传播方程在高代数中正规化并产生。拟议的计算模型可以充分提供梯度下降过程,并可能适合非常大规模稀少的NNPs,因为其表达性和培训效率很高,基于后向分析和计算图的NPs。