Despite impressive results, deep learning-based technologies also raise severe privacy and environmental concerns induced by the training procedure often conducted in data centers. In response, alternatives to centralized training such as Federated Learning (FL) have emerged. Perhaps unexpectedly, FL, in particular, is starting to be deployed at a global scale by companies that must adhere to new legal demands and policies originating from governments and the civil society for privacy protection. However, the potential environmental impact related to FL remains unclear and unexplored. This paper offers the first-ever systematic study of the carbon footprint of FL. First, we propose a rigorous model to quantify the carbon footprint, hence facilitating the investigation of the relationship between FL design and carbon emissions. Then, we compare the carbon footprint of FL to traditional centralized learning. Our findings show FL, despite being slower to converge, can be a greener technology than data center GPUs. Finally, we highlight and connect the reported results to the future challenges and trends in FL to reduce its environmental impact, including algorithms efficiency, hardware capabilities, and stronger industry transparency.
翻译:尽管取得了令人印象深刻的成果,深层次的学习技术也引起了数据中心经常开展的培训程序引起的严重的隐私和环境关切。作为回应,出现了联邦学习联合会(FL)等集中培训的替代办法。也许出乎意料的是,FL公司开始在全球范围部署,这些公司必须坚持来自政府和民间社会的新的保护隐私的法律要求和政策。然而,与FL公司相关的潜在环境影响仍然不明确且尚未探讨。本文件首次对FL公司的碳足迹进行了系统研究。首先,我们提出了一个严格的碳足迹量化模型,从而便利FL公司设计与碳排放之间的关系调查。然后,我们将FL公司的碳足迹与传统的集中学习进行比较。我们的研究结果显示,FL公司尽管比较缓慢地聚集起来,但可以成为比数据中心GPU公司更绿色的技术。最后,我们强调并连接所报告的结果与FL公司未来的挑战和趋势,以减少其环境影响,包括算法效率、硬件能力和更强大的工业透明度。