Modern AI practices all strive towards the same goal: better results. In the context of deep learning, the term "results" often refers to the achieved accuracy on a competitive problem set. In this paper, we adopt an idea from the emerging field of Green AI to consider energy consumption as a metric of equal importance to accuracy and to reduce any irrelevant tasks or energy usage. We examine the training stage of the deep learning pipeline from a sustainability perspective, through the study of hyperparameter tuning strategies and the model complexity, two factors vastly impacting the overall pipeline's energy consumption. First, we investigate the effectiveness of grid search, random search and Bayesian optimisation during hyperparameter tuning, and we find that Bayesian optimisation significantly dominates the other strategies. Furthermore, we analyse the architecture of convolutional neural networks with the energy consumption of three prominent layer types: convolutional, linear and ReLU layers. The results show that convolutional layers are the most computationally expensive by a strong margin. Additionally, we observe diminishing returns in accuracy for more energy-hungry models. The overall energy consumption of training can be halved by reducing the network complexity. In conclusion, we highlight innovative and promising energy-efficient practices for training deep learning models. To expand the application of Green AI, we advocate for a shift in the design of deep learning models, by considering the trade-off between energy efficiency and accuracy.
翻译:现代人工智能实践都追求同一个目标:更好的结果。在深度学习领域,"结果"这个词经常指的是在一个竞争的问题集上达到的准确度。在本文中,我们采用新兴领域--绿色AI的想法,将能源消耗作为一个同等重要的评估指标,以及减少任何无关的任务或能源消耗。我们从可持续性的角度审视深度学习管道的训练阶段,通过研究超参数调整策略和模型复杂性这两个极大影响整个流程能源消耗的因素。首先,我们研究了网格搜索、随机搜索和贝叶斯优化在超参数调整中的有效性,发现贝叶斯优化显著优于其他策略。此外,我们还分析了卷积神经网络结构中三种突出的层类型(卷积层、线性层和ReLU层)的能源消耗。结果显示,卷积层是计算开销最大的。此外,我们发现更具能量需求的模型带来的准确度增益出现递减。通过减少网络复杂度,训练的总能源消耗可以减少一半。总体而言,我们强调了训练深度学习模型的创新和有前途的节能实践。为了扩大绿色AI的应用,我们主张通过考虑能源效率和准确度之间的权衡来转变深度学习模型的设计。