Agentic systems offer a potential path to solve complex clinical tasks through collaboration among specialized agents, augmented by tool use and external knowledge bases. Nevertheless, for chest X-ray (CXR) interpretation, prevailing methods remain limited: (i) reasoning is frequently neither clinically interpretable nor aligned with guidelines, reflecting mere aggregation of tool outputs; (ii) multimodal evidence is insufficiently fused, yielding text-only rationales that are not visually grounded; and (iii) systems rarely detect or resolve cross-tool inconsistencies and provide no principled verification mechanisms. To bridge the above gaps, we present RadAgents, a multi-agent framework that couples clinical priors with task-aware multimodal reasoning and encodes a radiologist-style workflow into a modular, auditable pipeline. In addition, we integrate grounding and multimodal retrieval-augmentation to verify and resolve context conflicts, resulting in outputs that are more reliable, transparent, and consistent with clinical practice.
翻译:智能体系统通过专业化智能体间的协作,结合工具调用与外部知识库增强,为解决复杂临床任务提供了潜在路径。然而,当前胸部X光片(CXR)解读方法仍存在局限:(i)推理过程常缺乏临床可解释性且未遵循指南,仅体现为工具输出的简单聚合;(ii)多模态证据融合不足,仅生成缺乏视觉依据的纯文本解释;(iii)系统鲜少检测或解决跨工具不一致性问题,且缺乏原则性验证机制。为弥合上述差距,我们提出RadAgents——一个融合临床先验知识与任务感知多模态推理的多智能体框架,并将放射科医师式工作流程编码为模块化、可审计的管线。此外,我们集成基于多模态检索增强的语义落地机制,以验证并解决上下文冲突,从而生成更可靠、透明且符合临床实践的输出结果。