Deep Neural Networks (DNNs) are very popular because of their high performance in various cognitive tasks in Machine Learning (ML). Recent advancements in DNNs have brought beyond human accuracy in many tasks, but at the cost of high computational complexity. To enable efficient execution of DNN inference, more and more research works, therefore, exploit the inherent error resilience of DNNs and employ Approximate Computing (AC) principles to address the elevated energy demands of DNN accelerators. This article provides a comprehensive survey and analysis of hardware approximation techniques for DNN accelerators. First, we analyze the state of the art and by identifying approximation families, we cluster the respective works with respect to the approximation type. Next, we analyze the complexity of the performed evaluations (with respect to the dataset and DNN size) to assess the efficiency, the potential, and limitations of approximate DNN accelerators. Moreover, a broad discussion is provided, regarding error metrics that are more suitable for designing approximate units for DNN accelerators as well as accuracy recovery approaches that are tailored to DNN inference. Finally, we present how Approximate Computing for DNN accelerators can go beyond energy efficiency and address reliability and security issues, as well.
翻译:深神经网络(DNN)非常受欢迎,因为它们在机器学习(ML)中的各种认知任务中表现很高。 DNN最近的进步使许多任务超出了人的准确性,但代价是高计算复杂性。因此,为了能够高效率地执行DNN的推断,越来越多的研究工作利用DNN的内在误差复原力,并采用近似计算机(AC)原则来解决DNN加速器的高能量需求。这篇文章为DNN加速器提供了对硬件近似技术的全面调查和分析。首先,我们分析艺术状况,并通过确定近似型号,将相关工作归为近似型号。接下来,我们分析所进行的评估的复杂性(关于数据集和DNNN的大小),以评估近似DNN加速器的效率、潜力和局限性。此外,还广泛讨论了更适合DNNN的近似加速器设计近似单位的误差度度度测量方法,以及符合DNNN的精确度回收方法。最后,我们介绍了已进行的评估(关于数据集和DNNNNP的可靠性,以及D的计算机效率问题。