The latest work on language representations carefully integrates contextualized features into language model training, which enables a series of success especially in various machine reading comprehension and natural language inference tasks. However, the existing language representation models including ELMo, GPT and BERT only exploit plain context-sensitive features such as character or word embeddings. They rarely consider incorporating structured semantic information which can provide rich semantics for language representation. To promote natural language understanding, we propose to incorporate explicit contextual semantics from pre-trained semantic role labeling, and introduce an improved language representation model, Semantics-aware BERT (SemBERT), which is capable of explicitly absorbing contextual semantics over a BERT backbone. SemBERT keeps the convenient usability of its BERT precursor in a light fine-tuning way without substantial task-specific modifications. Compared with BERT, semantics-aware BERT is as simple in concept but more powerful. It obtains new state-of-the-art or substantially improves results on ten reading comprehension and language inference tasks.
翻译:语言代表的最新工作仔细地将背景特征纳入语言模式培训,从而能够取得一系列成功,特别是在各种机器阅读理解和自然语言推断任务方面。然而,现有的语言代表模式,包括ELMO、GPT和BERT, 仅仅利用了简单的背景敏感特征,如字符或字嵌入等。它们很少考虑纳入结构化的语义信息,为语言代表提供丰富的语义。为了促进自然语言理解,我们提议从培训前的语义角色标签中引入明确的背景语义,并引入一个改进的语言代表模式,即SemBERT(SemBERT),它能够明确地将背景语义吸收到BERT的骨干上。SemBERT保持其先质在轻巧调整时的易用性,而没有实质性的具体任务修改。与BERT相比,语义意识BERT在概念上非常简单,但更有力。它获得了新的状态或大幅度改进了十种阅读理解和语言推论任务的结果。