One of the most remarkable properties of word embeddings is the fact that they capture certain types of semantic and syntactic relationships. Recently, pre-trained language models such as BERT have achieved groundbreaking results across a wide range of Natural Language Processing tasks. However, it is unclear to what extent such models capture relational knowledge beyond what is already captured by standard word embeddings. To explore this question, we propose a methodology for distilling relational knowledge from a pre-trained language model. Starting from a few seed instances of a given relation, we first use a large text corpus to find sentences that are likely to express this relation. We then use a subset of these extracted sentences as templates. Finally, we fine-tune a language model to predict whether a given word pair is likely to be an instance of some relation, when given an instantiated template for that relation as input.
翻译:语言嵌入最显著的特性之一是它们捕捉到某些类型的语义和合成关系。最近,诸如BERT等经过预先训练的语言模型在一系列广泛的自然语言处理任务中取得了突破性成果。然而,尚不清楚这些模型在多大程度上捕捉到超越标准词嵌入所捕捉到的关联知识。为探讨这一问题,我们提出了一个方法,从经过训练的语文模型中提取关系知识。从某个特定关系的几个原始例子开始,我们首先使用一个大文本堆来找到可能表达这种关系的句子。我们然后使用这些提取的句子中的一组作为模板。最后,我们微调一个语言模型来预测一个特定词配对是否可能是某种关系的例子,当给该关系提供作为投入的瞬间模板的时候。