Pre-trained language models (PLMs) often take advantage of the monolingual and multilingual dataset that is freely available online to acquire general or mixed domain knowledge before deployment into specific tasks. Extra-large PLMs (xLPLMs) are proposed very recently to claim supreme performances over smaller-sized PLMs such as in machine translation (MT) tasks. These xLPLMs include Meta-AI's wmt21-dense-24-wide-en-X (2021) and NLLB (2022). In this work, we examine if xLPLMs are absolutely superior to smaller-sized PLMs in fine-tuning toward domain-specific MTs. We use two different in-domain data of different sizes: commercial automotive in-house data and clinical shared task data from the ClinSpEn2022 challenge at WMT2022. We choose popular Marian Helsinki as smaller sized PLM and two massive-sized Mega-Transformers from Meta-AI as xLPLMs. Our experimental investigation shows that 1) on smaller-sized in-domain commercial automotive data, xLPLM wmt21-dense-24-wide-en-X indeed shows much better evaluation scores using SacreBLEU and hLEPOR metrics than smaller-sized Marian, even though its score increase rate is lower than Marian after fine-tuning; 2) on relatively larger-size well prepared clinical data fine-tuning, the xLPLM NLLB tends to lose its advantage over smaller-sized Marian on two sub-tasks (clinical terms and ontology concepts) using ClinSpEn offered metrics METEOR, COMET, and ROUGE-L, and totally lost to Marian on Task-1 (clinical cases) on all official metrics including SacreBLEU and BLEU; 3) metrics do not always agree with each other on the same tasks using the same model outputs; 4) our clinical-Marian ranked No.1 using official metric SacreBLEU on Task-1 out of all teams.
翻译:培训前语言模型(PLM)通常利用单语言和多语言的在线数据集(PLM)来获取通用或混合域知识。 最近,我们提议使用超大型的PLM(xLPLM)来声称在机器翻译(MT)任务等规模较小的PLM(MT)任务中,超大型的PLM(xLPM),这些xLPLM(MetA-AI wmt21-dense-24-全部X(2021年)和NLLLLB(2022年)。在这项工作中,我们检查在对特定域MTMT的微调中,xLPLM(M)是否绝对优于小的PLM(M),我们使用两个更小的MALM-M(ROM-M-ML)小的硬性能数据模型,我们选择马里亚赫尔辛基作为所有小型的PLMM(MT-IL-I(NL-IL),我们的实验性研究显示在小型商业汽车上规模较小型的超小型的SUMLM(x)数据数据数据数据数据数据数据, 包括SLM-RM-NLODRM-NLM-NLM), 和S-NLM-S-S-S-S-C-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-SDRVDRVDRVDRVT-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-T-S-S-S-S-S-S-S-S-S-T-T-T-S-T-T-T-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-T-S-S-S-