Gradient descent optimizations and backpropagation are the most common methods for training neural networks, but they are computationally expensive for real time applications, need high memory resources, and are difficult to converge for many networks and large datasets. [Pseudo]inverse models for training neural network have emerged as powerful tools to overcome these issues. In order to effectively implement these methods, structured pruning maybe be applied to produce sparse neural networks. Although sparse neural networks are efficient in memory usage, most of their algorithms use the same fully loaded matrix calculation methods which are not efficient for sparse matrices. Tridiagonal matrices are one of the frequently used candidates for structuring neural networks, but they are not flexible enough to handle underfitting and overfitting problems as well as generalization properties. In this paper, we introduce a nonsymmetric, tridiagonal matrix with offdiagonal sparse entries and offset sub and super-diagonals as well algorithms for its [pseudo]inverse and determinant calculations. Traditional algorithms for matrix calculations, specifically inversion and determinant, of these forms are not efficient specially for large matrices, e.g. larger datasets or deeper networks. A decomposition for lower triangular matrices is developed and the original matrix is factorized into a set of matrices where their inverse matrices are calculated. For the cases where the matrix inverse does not exist, a least square type pseudoinverse is provided. The present method is a direct routine, i.e., executes in a predictable number of operations which is tested for randomly generated matrices with varying size. The results show significant improvement in computational costs specially when the size of matrix increases.
翻译:严重下沉优化和反向偏移是培训神经网络的最常见方法,但对于实时应用而言,它们计算成本昂贵,需要高记忆资源,而且对于许多网络和大型数据集来说难以趋同。 [Pseudo] 用于培训神经网络的反向模型已成为克服这些问题的强大工具。为了有效实施这些方法,结构化的调整或许可以用于产生稀薄的神经网络。虽然稀疏的神经网络在记忆使用方面是有效的,但它们的算法大多使用对稀薄的矩阵效率不高的完全装入的矩阵计算方法。三角形矩阵是经常用于构建神经网络的候选者之一,但是它们没有足够灵活地处理不完善和过度的问题以及概括性特性。在本文中,我们引入了一种非对称、三角矩阵的三角矩阵,并抵消了子和超直径矩阵的算法,对于稀释的矩阵计算方法效率不高。这些表格的传统的矩阵计算法,特别是用于构建神经网络的转换和决定性,这些表格的原始矩阵的计算结果不是更精确的。