Semantic Role Labeling (SRL) is believed to be a crucial step towards natural language understanding and has been widely studied. Recent years, end-to-end SRL with recurrent neural networks (RNN) has gained increasing attention. However, it remains a major challenge for RNNs to handle structural information and long range dependencies. In this paper, we present a simple and effective architecture for SRL which aims to address these problems. Our model is based on self-attention which can directly capture the relationships between two tokens regardless of their distance. Our single model achieves F$_1=83.4$ on the CoNLL-2005 shared task dataset and F$_1=82.7$ on the CoNLL-2012 shared task dataset, which outperforms the previous state-of-the-art results by $1.8$ and $1.0$ F$_1$ score respectively. Besides, our model is computationally efficient, and the parsing speed is 50K tokens per second on a single Titan X GPU.
翻译:据认为,语义作用标签(SRL)被认为是通向自然语言理解的关键一步,并且已经进行了广泛的研究。近年来,与经常性神经网络(RNN)的终端到终端SRL越来越受到关注。然而,对于RNN公司来说,处理结构性信息和长期依赖性仍然是一个重大挑战。在本文中,我们为SRL提出了一个旨在解决这些问题的简单而有效的架构。我们的模型基于自我关注,它可以直接捕捉两个象征之间的关系,而不论其距离如何。我们的单一模型在CONLL-2005共享任务数据集上达到了F$_1=83.4美元,而在CoNLL-2012共享任务数据集上达到了F_1=82.7美元,这分别比以往最新的结果高出1.8美元和1.0美元。此外,我们的模型是计算效率高的,而计算速度是单个泰坦X GPU每秒50K。