Human intelligence has the remarkable ability to assemble basic skills into complex ones so as to solve complex tasks. This ability is equally important for Artificial Intelligence (AI), and thus, we assert that in addition to the development of large, comprehensive intelligent models, it is equally crucial to equip such models with the capability to harness various domain-specific expert models for complex task-solving in the pursuit of Artificial General Intelligence (AGI). Recent developments in Large Language Models (LLMs) have demonstrated remarkable learning and reasoning abilities, making them promising as a controller to select, synthesize, and execute external models to solve complex tasks. In this project, we develop OpenAGI, an open-source AGI research platform, specifically designed to offer complex, multi-step tasks and accompanied by task-specific datasets, evaluation metrics, and a diverse range of extensible models. OpenAGI formulates complex tasks as natural language queries, serving as input to the LLM. The LLM subsequently selects, synthesizes, and executes models provided by OpenAGI to address the task. Furthermore, we propose a Reinforcement Learning from Task Feedback (RLTF) mechanism, which uses the task-solving result as feedback to improve the LLM's task-solving ability. Thus, the LLM is responsible for synthesizing various external models for solving complex tasks, while RLTF provides feedback to improve its task-solving ability, enabling a feedback loop for self-improving AI. We believe that the paradigm of LLMs operating various expert models for complex task-solving is a promising approach towards AGI. To facilitate the community's long-term improvement and evaluation of AGI's ability, we open-source the code, benchmark, and evaluation methods of the OpenAGI project at https://github.com/agiresearch/OpenAGI.
翻译:人类智慧有着将基本技能组合成复杂技能的非凡能力,以此来解决复杂任务。这种能力对于人工智能(AI)同样重要,因此,我们认为除了开发大型、全面的智能模型外,还要将这些模型配备具备各种领域专家模型的能力,以解决复杂任务,以期实现人工智能的通用性(AGI)。最近,大型语言模型(LLMs)在学习和推理方面表现出了惊人的能力,使其作为控制器选择、合成和执行外部模型来解决复杂任务的可行性变得更加具有前景。在本项目中,我们开发了OpenAGI,一个面向开放源代码的AGI研究平台,特别设计为提供复杂的多步骤任务、配备任务特定的数据集、评估指标和各种可扩展模型。 OpenAGI将复杂的任务公式化为自然语言查询,作为LLM的输入。 LLM随后选择、合成和执行由OpenAGI提供的模型,以解决任务。此外,我们提出了一种基于任务反馈的强化学习(RLTF)机制,它使用任务解决结果作为反馈来提高LLM的任务解决能力。因此,LLM负责合成各种用于解决复杂任务的外部模型,而RLTF提供反馈以提高其任务解决能力,从而实现自我完善人工智能的反馈循环。 我们认为,LLM操作各种专家模型以解决复杂任务的范式是实现AGI的一种有前途的方法。为了方便社区长期改善和评估AGI的能力,我们在https://github.com/agiresearch/OpenAGI上开源了OpenAGI项目的代码、基准测试和评估方法。