Recent advances in large-scale pre-training provide large models with the potential to learn knowledge from the raw text. It is thus natural to ask whether it is possible to leverage these large models as knowledge bases for downstream tasks. In this work, we answer the aforementioned question in unsupervised knowledge-grounded conversation. We explore various methods that best elicit knowledge from large models. Our human study indicates that, though hallucinations exist, large models post the unique advantage of being able to output common sense and summarize facts that cannot be directly retrieved from the search engine. To better exploit such generated knowledge in dialogue generation, we treat the generated knowledge as a noisy knowledge source and propose the posterior-based reweighing as well as the noisy training strategy. Empirical results on two benchmarks show advantages over the state-of-the-art methods.
翻译:大规模培训前的近期进展为大型模型提供了从原始文本中学习知识的潜力,因此,很自然地会问是否可以利用这些大型模型作为下游任务的知识基础。在这项工作中,我们在未经监督的知识基础上对话中回答上述问题。我们探索了从大型模型中获取知识的最佳方法。我们的人类研究表明,尽管存在幻觉,大型模型具有能够产生常识的独特优势,并总结无法直接从搜索引擎中获取的事实。为了在对话中更好地利用这些生成的知识,我们把产生的知识当作一个吵闹的知识来源,并提出以远地点为基础的重新拉动和吵闹式培训战略。两个基准的实证结果表明,最先进的方法具有优势。