Neural networks (NNs) are making a large impact both on research and industry. Nevertheless, as NNs' accuracy increases, it is followed by an expansion in their size, required number of compute operations and energy consumption. Increase in resource consumption results in NNs' reduced adoption rate and real-world deployment impracticality. Therefore, NNs need to be compressed to make them available to a wider audience and at the same time decrease their runtime costs. In this work, we approach this challenge from a causal inference perspective, and we propose a scoring mechanism to facilitate structured pruning of NNs. The approach is based on measuring mutual information under a maximum entropy perturbation, sequentially propagated through the NN. We demonstrate the method's performance on two datasets and various NNs' sizes, and we show that our approach achieves competitive performance under challenging conditions.
翻译:神经网络(NNs)正在对研究和工业产生巨大影响。然而,随着NNs精确度的提高,随着NNs精确度的提高,随之而来的是其规模、计算操作和能源消耗量的扩大,计算作业和能源消耗量的所需数量。资源消耗量的增加导致NNs采用率下降,实际部署不切实际。因此,NNS需要压缩,以便将其提供给更广泛的受众,同时降低运行时间成本。在这项工作中,我们从因果关系的角度来应对这一挑战,我们建议一个评分机制,以便利对NNS进行结构化的排查。这个方法的基础是在最大恒温渗透状态下衡量相互信息,通过NNs按顺序传播。我们展示了这种方法在两个数据集和各种NNs规模上的表现,我们表明我们的方法在挑战的条件下取得了竞争性的业绩。