Deep learning using neural networks is an effective technique for generating models of complex data. However, training such models can be expensive when networks have large model capacity resulting from a large number of layers and nodes. For training in such a computationally prohibitive regime, dimensionality reduction techniques ease the computational burden, and allow implementations of more robust networks. We propose a novel type of such dimensionality reduction via a new deep learning architecture based on fast matrix multiplication of a Kronecker product decomposition; in particular our network construction can be viewed as a Kronecker product-induced sparsification of an "extended" fully connected network. Analysis and practical examples show that this architecture allows a neural network to be trained and implemented with a significant reduction in computational time and resources, while achieving a similar error level compared to a traditional feedforward neural network.
翻译:利用神经网络进行深思熟虑是生成复杂数据模型的有效技术,然而,当网络具有大量层次和节点产生的巨大模型能力时,培训这类模型的费用可能非常昂贵。对于这种计算上令人望而生畏的系统的培训来说,多维性减少技术可以减轻计算负担,并能够实施更强大的网络。我们建议通过基于克罗内克产品分解快速增殖的新型深层学习结构来减少这种维度;特别是,我们的网络建设可被视为Kronecker产品引发的“扩展”完全连通网络的封闭。分析和实际实例表明,这一结构允许对神经网络进行培训和实施,同时大大减少计算时间和资源,同时实现与传统的向前神经网络类似的错误水平。