Neural networks have achieved tremendous success in a large variety of applications. However, their memory footprint and computational demand can render them impractical in application settings with limited hardware or energy resources. In this work, we propose a novel algorithm to find efficient low-rank subnetworks. Remarkably, these subnetworks are determined and adapted already during the training phase and the overall time and memory resources required by both training and evaluating them are significantly reduced. The main idea is to restrict the weight matrices to a low-rank manifold and to update the low-rank factors rather than the full matrix during training. To derive training updates that are restricted to the prescribed manifold, we employ techniques from dynamic model order reduction for matrix differential equations. This allows us to provide approximation, stability, and descent guarantees. Moreover, our method automatically and dynamically adapts the ranks during training to achieve the desired approximation accuracy. The efficiency of the proposed method is demonstrated through a variety of numerical experiments on fully-connected and convolutional networks.
翻译:神经网络在各种各样的应用中取得了巨大的成功,然而,它们的记忆足迹和计算需求使得它们在硬件或能源资源有限的应用环境中变得不切实际。在这项工作中,我们建议一种新型算法,以找到高效的低级别子网络。值得注意的是,这些子网络在培训阶段已经确定和调整,培训和评价这两个网络所需的全部时间和记忆资源也大大减少。主要想法是将体重矩阵限制在低级别,更新低级别因素,而不是培训期间的全部矩阵。为了获得限于规定的多功能的培训更新,我们采用了矩阵差异方程式动态模式减少命令的技术。这使我们能够提供近似、稳定性和血统保证。此外,我们的方法在培训期间自动和动态地调整队伍,以达到理想的近似精度。通过对完全连接的和动态的网络进行各种数字实验,可以证明拟议方法的效率。