In recent years, larger and deeper models are springing up and continuously pushing state-of-the-art (SOTA) results across various fields like natural language processing (NLP) and computer vision (CV). However, despite promising results, it needs to be noted that the computations required by SOTA models have been increased at an exponential rate. Massive computations not only have a surprisingly large carbon footprint but also have negative effects on research inclusiveness and deployment on real-world applications. Green deep learning is an increasingly hot research field that appeals to researchers to pay attention to energy usage and carbon emission during model training and inference. The target is to yield novel results with lightweight and efficient technologies. Many technologies can be used to achieve this goal, like model compression and knowledge distillation. This paper focuses on presenting a systematic review of the development of Green deep learning technologies. We classify these approaches into four categories: (1) compact networks, (2) energy-efficient training strategies, (3) energy-efficient inference approaches, and (4) efficient data usage. For each category, we discuss the progress that has been achieved and the unresolved challenges.
翻译:近年来,较大和更深层次的模型正在自然语言处理和计算机愿景(CV)等各个领域涌现并不断推进最新科技成果。然而,尽管取得了令人乐观的成果,但需要指出的是,SOTA模型所要求的计算以指数速度增长。大规模计算不仅具有令人惊讶的巨大碳足迹,而且对研究包容性和现实世界应用的部署产生了负面影响。绿色深层学习是一个日益热门的研究领域,需要研究人员在模型培训和推论期间关注能源使用和碳排放。目标是以轻量和高效技术产生新的结果。许多技术可以用于实现这一目标,如模型压缩和知识蒸馏。本文的重点是系统地审查绿色深层学习技术的发展。我们将这些方法分为四类:(1) 小型网络,(2) 节能培训战略,(3) 节能推断方法,以及(4) 高效的数据使用。我们讨论每一类中已经取得的进展和尚未解决的挑战。