Recent works show that discourse analysis benefits from modeling intra- and inter-sentential levels separately, where proper representations for text units of different granularities are desired to capture both the meaning of text units and their relations to the context. In this paper, we propose to take advantage of transformers to encode contextualized representations of units of different levels to dynamically capture the information required for discourse dependency analysis on intra- and inter-sentential levels. Motivated by the observation of writing patterns commonly shared across articles, we propose a novel method that treats discourse relation identification as a sequence labelling task, which takes advantage of structural information from the context of extracted discourse trees, and substantially outperforms traditional direct-classification methods. Experiments show that our model achieves state-of-the-art results on both English and Chinese datasets.
翻译:最近的工作表明,话语分析从分别建模内和文室间层次的模型中获益,在这种模型中,希望对不同颗粒的文字单位进行适当的表述,既反映文字单位的含义,又反映它们与上下文的关系。在本文件中,我们提议利用变压器将不同层次单位的背景化表述编码,动态地捕捉到对文系内和文系间层次的谈话依赖性分析所需的信息。我们以观察不同文章之间共同共有的写作模式为动力,提出一种新的方法,将话语系关系识别视为一个顺序标签任务,利用从提取的讲义树背景中得出的结构信息,大大超越传统的直接分类方法。实验表明,我们的模型在英文和中文数据集中都取得了最新的结果。