Pre-trained Language Models (PLMs) have shown superior performance on various downstream Natural Language Processing (NLP) tasks. However, conventional pre-training objectives do not explicitly model relational facts in text, which are crucial for textual understanding. To address this issue, we propose a novel contrastive learning framework ERICA to obtain a deep understanding of the entities and their relations in text. Specifically, we define two novel pre-training tasks to better understand entities and relations: (1) the entity discrimination task to distinguish which tail entity can be inferred by the given head entity and relation; (2) the relation discrimination task to distinguish whether two relations are close or not semantically, which involves complex relational reasoning. Experimental results demonstrate that ERICA can improve typical PLMs (BERT and RoBERTa) on several language understanding tasks, including relation extraction, entity typing and question answering, especially under low-resource settings.
翻译:培训前语言模式(PLMs)在各种下游自然语言处理(NLP)任务方面表现优异,但是,传统的培训前目标没有明确地在文本中以关系事实为模范,这些事实对文字理解至关重要。为了解决这一问题,我们提出一个新的对比式学习框架ERICA,以深入了解实体及其在文本中的关系。具体地说,我们界定了两项新的培训前任务,以更好地了解实体和关系:(1) 实体歧视任务,以区分哪个尾实体可以被某个实体和关系推断为尾实体;(2) 区分两种关系是否密切,这涉及复杂的关系推理;实验结果表明,ERICA可以改进典型的PLMs(BERTA和ROBERTA),执行几种语言理解任务,包括提取关系、实体打字和回答问题,特别是在低资源环境下。