In the last five years, the rise of the self-attentional Transformer-based architectures led to state-of-the-art performances over many natural language tasks. Although these approaches are increasingly popular, they require large amounts of data and computational resources. There is still a substantial need for benchmarking methodologies ever upwards on under-resourced languages in data-scarce application conditions. Most pre-trained language models were massively studied using the English language and only a few of them were evaluated on French. In this paper, we propose a unified benchmark, focused on evaluating models quality and their ecological impact on two well-known French spoken language understanding tasks. Especially we benchmark thirteen well-established Transformer-based models on the two available spoken language understanding tasks for French: MEDIA and ATIS-FR. Within this framework, we show that compact models can reach comparable results to bigger ones while their ecological impact is considerably lower. However, this assumption is nuanced and depends on the considered compression method.
翻译:在过去5年中,基于自我注意的变异器结构的兴起导致在许多自然语言任务中出现最先进的表现。虽然这些方法越来越流行,但它们需要大量的数据和计算资源。在数据偏差的应用条件下,仍然非常需要为资源不足的语言制定基准方法。大多数预先培训的语言模式使用英语进行了大量研究,只有少数是用法语进行评估。在本文件中,我们提出了一个统一的基准,重点是评估模型质量及其对两种众所周知的法语口语理解任务产生的生态影响。特别是我们根据两种现有的法语口语理解任务:MEDIA和ATIS-FR,将13个成熟的变异器模型作为基准。在此框架内,我们表明,在生态影响大大降低的情况下,紧凑模式可以取得与较大语言相似的结果。然而,这一假设是细微的,取决于经过考虑的压缩方法。