A key missing ability of current language models (LMs) is grounding to real-world environments. Most existing work for grounded language understanding uses LMs to directly generate plans that can be executed in the environment to achieve the desired effects. It casts the burden of ensuring grammaticality, faithfulness, and controllability all on the LMs. We propose Pangu, a generic framework for grounded language understanding that capitalizes on the discriminative ability of LMs instead of their generative ability. Pangu consists of a symbolic agent and a neural LM working in a concerted fashion: the agent explores the environment to incrementally construct valid candidate plans, and the LM evaluates the plausibility of the candidate plans to guide the search process. A case study on the challenging problem of knowledge base question answering (KBQA), which features a massive environment, demonstrates the remarkable effectiveness and flexibility of Pangu: A BERT-base LM is sufficient for achieving a new state of the art on standard KBQA datasets, and larger LMs further improve the performance by a large margin. Pangu also enables, for the first time, effective few-shot in-context learning for KBQA with large LMs such as Codex.
翻译:目前语言模型缺少的关键能力(LMS)正在以现实世界环境为基础。目前大多数基于基础语言理解的工作都利用LMS直接制定可以在环境中执行的计划,以实现预期效果。它给LMS带来了确保语言学、忠诚和可控性的责任。我们提议Pangu,一个基于语言理解的一般性框架,利用LMS的歧视性能力,而不是基因化能力。Pangu由象征性的代理和神经力LMM组成,以协调一致的方式工作:代理商探索环境,逐步制定有效的候选计划,LM评估候选人计划在环境中指导搜索过程的可信度。关于具有挑战性的知识基础回答问题的案例研究(KBQA),其特点是大环境,显示了Pangu的显著效力和灵活性:BERT-B基地LMM足以实现标准的KBQA数据集的新状态,更大的LMSMS进一步大幅改进业绩。Pangu 第一次,Pangu还能够将有效的KACS-Chol 用于大型的KCON。