Recent Deep Neural Networks (DNNs) managed to deliver superhuman accuracy levels on many AI tasks. Several applications rely more and more on DNNs to deliver sophisticated services and DNN accelerators are becoming integral components of modern systems-on-chips. DNNs perform millions of arithmetic operations per inference and DNN accelerators integrate thousands of multiply-accumulate units leading to increased energy requirements. Approximate computing principles are employed to significantly lower the energy consumption of DNN accelerators at the cost of some accuracy loss. Nevertheless, recent research demonstrated that complex DNNs are increasingly sensitive to approximation. Hence, the obtained energy savings are often limited when targeting tight accuracy constraints. In this work, we present a dynamically configurable approximate multiplier that supports three operation modes, i.e., exact, positive error, and negative error. In addition, we propose a filter-oriented approximation method to map the weights to the appropriate modes of the approximate multiplier. Our mapping algorithm balances the positive with the negative errors due to the approximate multiplications, aiming at maximizing the energy reduction while minimizing the overall convolution error. We evaluate our approach on multiple DNNs and datasets against state-of-the-art approaches, where our method achieves 18.33% energy gains on average across 7 NNs on 4 different datasets for a maximum accuracy drop of only 1%.
翻译:最近深神经网络(DNN)设法在很多AI任务上提供了超人精度水平。一些应用程序越来越多地依靠DNN来提供尖端服务,DNN加速器正在成为现代系统-芯片中不可分割的组成部分。DNNS按每个推算进行数以百万计的算术操作,DNN加速器结合了数千个乘积单位,导致能源需求增加。采用了近似计算原则,以降低某些精度损失为代价,大幅降低DNN加速器的能源消耗量。然而,最近的研究表明,复杂的DNNN越来越敏感于近似。因此,在瞄准精确度限制时,获得的节能往往有限。在这项工作中,DNNNS使用一种动态可配置近似数的乘数乘数,支持三种操作模式,即精确、正误差和负差。此外,我们提出了一种面向过滤的近似近似近似近似方法,将DNNC加速器的重量与由于近似增加而出现的负差相平衡。我们用负差数计算算法的目的是最大限度地减少能源的精确度,同时最大限度地减少能源的精确度,同时尽量减少整个DNNNNR方法在18位上取得最大收益。我们18种方法上。我们18种平均的能量率方法。我们用在18种不同的标准中,我们用率方法上对18个的能量计算了。我们测算。我们测算。我们测算方法在18个平均的能量法中测取了一种不同的第一种不同的标准。我们测算方法。