Recent progress in pretraining language models on large textual corpora led to a surge of improvements for downstream NLP tasks. Whilst learning linguistic knowledge, these models may also be storing relational knowledge present in the training data, and may be able to answer queries structured as "fill-in-the-blank" cloze statements. Language models have many advantages over structured knowledge bases: they require no schema engineering, allow practitioners to query about an open class of relations, are easy to extend to more data, and require no human supervision to train. We present an in-depth analysis of the relational knowledge already present (without fine-tuning) in a wide range of state-of-the-art pretrained language models. We find that (i) without fine-tuning, BERT contains relational knowledge competitive with traditional NLP methods that have some access to oracle knowledge, (ii) BERT also does remarkably well on open-domain question answering against a supervised baseline, and (iii) certain types of factual knowledge are learned much more readily than others by standard language model pretraining approaches. The surprisingly strong ability of these models to recall factual knowledge without any fine-tuning demonstrates their potential as unsupervised open-domain QA systems. The code to reproduce our analysis is available at https://github.com/facebookresearch/LAMA.
翻译:在大型文本公司培训前语言模型方面最近的进展导致下游国家语言平台任务改进的激增。虽然学习语言知识,这些模型也可能储存培训数据中的现有关系知识,并可能回答以“填补空白”语团语团语团语团语团语团语团语团语团语团语团语团语团语团语团语团语团语组的质询。语言模型与结构化知识库相比有许多优势:它们不需要系统工程学,让从业人员就开放型关系类别进行问询,很容易扩展到更多的数据,也不需要人类监督来进行培训。我们深入分析了在一系列经过事先训练的通用语言模型中已经存在的(不作微调)的关系知识。我们发现,(一)不作微调,生物伦理中心(BERT)拥有与传统国家语言平台语团语团语团语团语言联盟(NLP)方法具有竞争性的关系知识,(二)它们不需要某种系统图案组式工程团语团知识,(BERT)在根据受监督的基线回答的公开式问题上也非常出色,以及(三)某些类型的事实知识比标准语言模型/培训前方法更容易学习。这些模型在回顾实际知识方面的能力惊人。在不进行精确分析时可以进行。