Human intelligence has the remarkable ability to assemble basic skills into complex ones so as to solve complex tasks. This ability is equally important for Artificial Intelligence (AI), and thus, we assert that in addition to the development of large, comprehensive intelligent models, it is equally crucial to equip such models with the capability to harness various domain-specific expert models for complex task-solving in the pursuit of Artificial General Intelligence (AGI). Recent developments in Large Language Models (LLMs) have demonstrated remarkable learning and reasoning abilities, making them promising as a controller to select, synthesize, and execute external models to solve complex tasks. In this project, we develop OpenAGI, an open-source AGI research platform, specifically designed to offer complex, multi-step tasks and accompanied by task-specific datasets, evaluation metrics, and a diverse range of extensible models. OpenAGI formulates complex tasks as natural language queries, serving as input to the LLM. The LLM subsequently selects, synthesizes, and executes models provided by OpenAGI to address the task. Furthermore, we propose a Reinforcement Learning from Task Feedback (RLTF) mechanism, which uses the task-solving result as feedback to improve the LLM's task-solving ability. Thus, the LLM is responsible for synthesizing various external models for solving complex tasks, while RLTF provides feedback to improve its task-solving ability, enabling a feedback loop for self-improving AI. We believe that the paradigm of LLMs operating various expert models for complex task-solving is a promising approach towards AGI. To facilitate the community's long-term improvement and evaluation of AGI's ability, we open-source the code, benchmark, and evaluation methods of the OpenAGI project at https://github.com/agiresearch/OpenAGI.
翻译:人类智能具有将基础技能组合成复杂技能以解决复杂任务的显著能力。对于人工智能(AI)而言,该能力同样重要。因此,我们认为除了开发大型、全面的智能模型外,还同样关键的是赋予这些模型与各种领域特定专家模型配合协同解决复杂任务的能力,以追求人工通用智能(AGI)。最近大型语言模型(LLM)的发展表明,它们具有出色的学习和推理能力,因而被视为有希望作为控制器来选择、综合和执行外部模型以解决复杂任务。在本项目中,我们开发了 OpenAGI,这是一个开源的 AGI 研究平台,专门设计用于提供复杂的多步骤任务,附带任务特定数据集、评估指标和各种可扩展模型。OpenAGI 将复杂任务构建为自然语言查询,作为 LLM 的输入。LLM 随后选择、综合和执行 OpenAGI 提供的模型以解决任务。此外,我们提出了一种基于任务反馈的强化学习(RLTF)机制,该机制使用任务解决结果作为反馈,以提高 LLM 的任务解决能力。因此,LLM 负责合成各种外部模型,以解决复杂任务,而 RLTF 提供反馈以改善其任务解决能力,为自我改进的 AI 提供反馈循环。我们相信,LLMs 操作各种专家模型解决复杂任务的范式是实现 AGI 的有前途的方法。为了方便社区对 AGI 能力的长期改进和评估,我们在 https://github.com/agiresearch/OpenAGI 开源了 OpenAGI 项目的代码、基准测试和评估方法。