Recent advances in spoken language understanding benefited from Self-Supervised models trained on large speech corpora. For French, the LeBenchmark project has made such models available and has led to impressive progress on several tasks including spoken language understanding. These advances have a non-negligible cost in terms of computation time and energy consumption. In this paper, we compare several learning strategies trying to reduce such cost while keeping competitive performance. At the same time we propose an extensive analysis where we measure the cost of our models in terms of training time and electric energy consumption, hopefully promoting a comprehensive evaluation procedure. The experiments are performed on the FSC and MEDIA corpora, and show that it is possible to reduce the learning cost while maintaining state-of-the-art performance and using SSL models.
翻译:在口头语言理解方面最近的进展得益于在大型语言公司方面培训的自我监督模型,在法语方面,LeBenchmark项目提供了这种模型,并导致在包括口头语言理解在内的若干任务方面取得令人印象深刻的进展,这些进步在计算时间和能源消耗方面成本不可忽略,在本文件中,我们比较了一些学习战略,力求降低这种成本,同时保持有竞争力的绩效。同时,我们提议进行广泛的分析,从培训时间和电力消耗方面衡量我们模型的成本,希望促进全面的评估程序。这些实验是在FSC和MEDIA公司进行的,表明在保持最新业绩和使用SSL模式的同时,可以降低学习成本。