The development of intelligent agents, particularly those powered by language models (LMs), has shown a critical role in various environments that require intelligent and autonomous decision-making. Environments are not passive testing grounds, and they represent the data required for agents to learn and exhibit in very challenging conditions that require adaptive, complex, and autonomous capacity to make decisions. While the paradigm of scaling models and datasets has led to remarkable emergent capabilities, we argue that scaling the structure, fidelity, and logical consistency of agent reasoning within these environments is a crucial, yet underexplored, dimension of AI research. This paper introduces a neuro-symbolic multi-agent architecture where the belief states of individual agents are formally represented as Kripke models. This foundational choice enables them to reason about known concepts of \emph{possibility} and \emph{necessity} using the formal language of modal logic. In this work, we use immutable, domain-specific knowledge to make an informed root cause diagnosis, which is encoded as logical constraints essential for proper, reliable, and explainable diagnosis. In the proposed model, we show constraints that actively guide the hypothesis generation of LMs, effectively preventing them from reaching physically or logically untenable conclusions. In a high-fidelity simulated particle accelerator environment, our system successfully diagnoses complex, cascading failures by combining the powerful semantic intuition of LMs with the rigorous, verifiable validation of modal logic and a factual world model and showcasing a viable path toward more robust, reliable, and verifiable autonomous agents.
翻译:智能体,特别是由语言模型驱动的智能体,其发展在各种需要智能自主决策的环境中展现出关键作用。环境并非被动的测试场,它们代表了智能体在需要适应性、复杂性和自主决策能力的极具挑战性条件下进行学习和表现所需的数据。尽管扩大模型和数据规模的范式已催生出显著的新兴能力,但我们认为,在这些环境中扩展智能体推理的结构、保真度和逻辑一致性,是人工智能研究中一个至关重要却尚未被充分探索的维度。本文介绍了一种神经符号多智能体架构,其中单个智能体的信念状态被形式化地表示为克里普克模型。这一基础性选择使得它们能够使用模态逻辑的形式语言,对已知的"可能性"和"必然性"概念进行推理。在本工作中,我们利用不可变的领域特定知识进行有依据的根本原因诊断,这些知识被编码为逻辑约束,对于实现正确、可靠且可解释的诊断至关重要。在所提出的模型中,我们展示了能够主动引导语言模型进行假设生成的约束,有效防止其得出物理上或逻辑上站不住脚的结论。在一个高保真模拟粒子加速器环境中,我们的系统通过将语言模型强大的语义直觉与模态逻辑及事实世界模型的严格可验证性相结合,成功诊断了复杂的级联故障,展示了一条通往更鲁棒、可靠且可验证的自主智能体的可行路径。