Parallel cross-lingual summarization data is scarce, requiring models to better use the limited available cross-lingual resources. Existing methods to do so often adopt sequence-to-sequence networks with multi-task frameworks. Such approaches apply multiple decoders, each of which is utilized for a specific task. However, these independent decoders share no parameters, hence fail to capture the relationships between the discrete phrases of summaries in different languages, breaking the connections in order to transfer the knowledge of the high-resource languages to low-resource languages. To bridge these connections, we propose a novel Multi-Task framework for Cross-Lingual Abstractive Summarization (MCLAS) in a low-resource setting. Employing one unified decoder to generate the sequential concatenation of monolingual and cross-lingual summaries, MCLAS makes the monolingual summarization task a prerequisite of the cross-lingual summarization (CLS) task. In this way, the shared decoder learns interactions involving alignments and summary patterns across languages, which encourages attaining knowledge transfer. Experiments on two CLS datasets demonstrate that our model significantly outperforms three baseline models in both low-resource and full-dataset scenarios. Moreover, in-depth analysis on the generated summaries and attention heads verifies that interactions are learned well using MCLAS, which benefits the CLS task under limited parallel resources.
翻译:平行的跨语言汇总数据十分稀少,需要模型来更好地利用有限的现有跨语言资源。现有的方法往往采用多任务框架的顺序对顺序网络。这些方法采用多个解码器,每个解码器都用于具体任务。然而,这些独立的解码器没有共享参数,因此无法捕捉不同语言摘要中不同词句之间的关系,从而中断连接,将高资源语言知识传授给低资源语言。为了连接这些连接,我们提议了一个创新的跨语言摘要汇总多任务框架(多任务框架),在低资源环境下进行跨语言汇总。使用一个统一的解码器来产生单语和跨语言摘要的顺序组合。但是,这些独立解码器将单语拼拼拼拼拼任务作为跨语言拼凑任务的先决条件。在这种方式上,共享解码器学习了跨语言的校对和摘要模式的相互作用,鼓励实现知识转让。在两个CLSS摘要中进行实验表明,在C-LS的模型中,在深度分析中,在深度分析中,我们模型中,在深度的CLSB任务中,在深度分析中,在深度分析中,在深度分析中,核心任务模型中,对三个基准模型分析中,在深度分析中,在深度分析中,核心任务模型分析中,这些是大量。