Recent advances in the field of abstractive summarization leverage pre-trained language models rather than train a model from scratch. However, such models are sluggish to train and accompanied by a massive overhead. Researchers have proposed a few lightweight alternatives such as smaller adapters to mitigate the drawbacks. Nonetheless, it remains uncertain whether using adapters benefits the task of summarization, in terms of improved efficiency without an unpleasant sacrifice in performance. In this work, we carry out multifaceted investigations on fine-tuning and adapters for summarization tasks with varying complexity: language, domain, and task transfer. In our experiments, fine-tuning a pre-trained language model generally attains a better performance than using adapters; the performance gap positively correlates with the amount of training data used. Notably, adapters exceed fine-tuning under extremely low-resource conditions. We further provide insights on multilinguality, model convergence, and robustness, hoping to shed light on the pragmatic choice of fine-tuning or adapters in abstractive summarization.
翻译:然而,目前尚不确定的是,使用适应器是否有利于总结工作,即提高工作效率,而不令人不快地牺牲业绩。在这项工作中,我们进行了多方面的调查,对精细调整和调整工作进行了多方面的调查,以完成复杂程度不一的总结工作:语言、域和任务转移。在我们的实验中,对预先培训的语言模型进行微调和调整,通常比使用适应器取得更好的性能;绩效差距与使用的培训数据数量成正比。值得注意的是,适应器在极其低资源条件下超过了微调。我们进一步介绍了多语言性、模式趋同和稳健性,希望说明在抽象的总结中精密调整或调整的实用选择。