This survey reviews works in which language models (LMs) are augmented with reasoning skills and the ability to use tools. The former is defined as decomposing a potentially complex task into simpler subtasks while the latter consists in calling external modules such as a code interpreter. LMs can leverage these augmentations separately or in combination via heuristics, or learn to do so from demonstrations. While adhering to a standard missing tokens prediction objective, such augmented LMs can use various, possibly non-parametric external modules to expand their context processing ability, thus departing from the pure language modeling paradigm. We therefore refer to them as Augmented Language Models (ALMs). The missing token objective allows ALMs to learn to reason, use tools, and even act, while still performing standard natural language tasks and even outperforming most regular LMs on several benchmarks. In this work, after reviewing current advance in ALMs, we conclude that this new research direction has the potential to address common limitations of traditional LMs such as interpretability, consistency, and scalability issues.
翻译:本调查审查的工作是使语言模型(LMS)得到推理技能和使用工具的能力的增强,前者被定义为将潜在复杂的任务分解成更简单的子任务,而后者则包括调用诸如代码翻译员等外部模块。LMS可以分别或混合地利用这些增强功能,或从演示中学习这样做。这种增强的LMS在坚持标准缺失标志预测目标的同时,可以使用各种可能不是参数的外部模块来扩大其上下文处理能力,从而偏离纯语言模型。因此,我们称它们为强化语言模型(ALMs)。缺失的象征性目标使ALMs学会理性、使用工具甚至行动,同时仍然执行标准的自然语言任务,甚至在某些基准上超过最常规的LMs。在审查了ALMs当前的进步之后,我们得出结论,这一新的研究方向有可能解决传统LMs的共同局限性,例如可解释性、一致性和可缩缩缩略性问题。