In recent times, there has been definitive progress in the field of NLP, with its applications growing as the utility of our language models increases with advances in their performance. However, these models require a large amount of computational power and data to train, consequently leading to large carbon footprints. Therefore, it is imperative that we study the carbon efficiency and look for alternatives to reduce the overall environmental impact of training models, in particular large language models. In our work, we assess the performance of models for machine translation, across multiple language pairs to assess the difference in computational power required to train these models for each of these language pairs and examine the various components of these models to analyze aspects of our pipeline that can be optimized to reduce these carbon emissions.
翻译:近来,随着语言模型的使用随着性能的提高而增加,在NLP领域已经取得了明确的进展,其应用随着我们语言模型的使用量的增加而增长,然而,这些模型需要大量的计算力和数据来培训,从而导致巨大的碳足迹。因此,我们必须研究碳效率,寻找替代办法,以减少培训模型,特别是大型语言模型的总体环境影响。在我们的工作中,我们评估了多种语言对的机器翻译模型的性能,以评估为每种语言对培训这些模型所需的计算能力差异,并审查这些模型的各个组成部分,以分析我们为减少这些碳排放可以优化的管道的各个方面。