Recently, there has been a trend of shifting the execution of deep learning inference tasks toward the edge of the network, closer to the user, to reduce latency and preserve data privacy. At the same time, growing interest is being devoted to the energetic sustainability of machine learning. At the intersection of these trends, we hence find the energetic characterization of machine learning at the edge, which is attracting increasing attention. Unfortunately, calculating the energy consumption of a given neural network during inference is complicated by the heterogeneity of the possible underlying hardware implementation. In this work, we hence aim at profiling the energetic consumption of inference tasks for some modern edge nodes and deriving simple but realistic models. To this end, we performed a large number of experiments to collect the energy consumption of convolutional and fully connected layers on two well-known edge boards by NVIDIA, namely Jetson TX2 and Xavier. From the measurements, we have then distilled a simple, practical model that can provide an estimate of the energy consumption of a certain inference task on the considered boards. We believe that this model can be used in many contexts as, for instance, to guide the search for efficient architectures in Neural Architecture Search, as a heuristic in Neural Network pruning, or to find energy-efficient offloading strategies in a Split computing context, or simply to evaluate the energetic performance of Deep Neural Network architectures.
翻译:最近出现了一种趋势,即将深学习推论任务的执行转向网络边缘,更接近用户,以减少潜伏,并保护数据隐私。与此同时,人们越来越关注机器学习的活力可持续性。在这些趋势交汇处,我们发现在边缘对机器学习的强烈特征描述正在引起越来越多的注意。不幸的是,在推论期间计算特定神经网络的能量消耗由于可能的基本硬件实施过程的异质性而变得复杂。在这项工作中,我们的目标是分析一些现代边缘节点大量消耗推断任务并产生简单而现实的模型。为此,我们进行了大量实验,以收集VIVIDA的两个著名边缘板(即Jetson TX2和Xavier)的革命性和完全连接层的能源消耗。从测量中,我们随后重新确定了一个简单而实用的模型,可以提供对所考虑的板上某种精度任务的能量消耗的估计。我们认为,在很多背景下,可以使用这种模型来收集革命性与完全相连的层层的能源消耗量,可以用来作为智能的建筑结构,用来在结构上找到一个快速的搜索,作为实例。