Artificial Intelligence has achieved remarkable advancements in recent years, yet much of its progress relies on identifying increasingly complex correlations. Enabling causality awareness in AI has the potential to enhance its performance by enabling a deeper understanding of the underlying mechanisms of the environment. In this paper, we introduce DODO, an algorithm defining how an Agent can autonomously learn the causal structure of its environment through repeated interventions. We assume a scenario where an Agent interacts with a world governed by a causal Directed Acyclic Graph (DAG), which dictates the system's dynamics but remains hidden from the Agent. The Agent's task is to accurately infer the causal DAG, even in the presence of noise. To achieve this, the Agent performs interventions, leveraging causal inference techniques to analyze the statistical significance of observed changes. Results show better performance for DODO, compared to observational approaches, in all but the most limited resource conditions. DODO is often able to reconstruct with as low as zero errors the structure of the causal graph. In the most challenging configuration, DODO outperforms the best baseline by +0.25 F1 points.
翻译:近年来,人工智能取得了显著进展,但其大部分进步依赖于识别日益复杂的相关性。通过使人工智能具备因果感知能力,可以使其更深入地理解环境的内在机制,从而提升其性能。本文提出DODO算法,该算法定义了智能体如何通过重复干预自主学习其环境的因果结构。我们假设存在一个场景,其中智能体与一个由因果有向无环图(DAG)支配的世界进行交互,该图决定了系统的动态特性,但对智能体保持隐藏。智能体的任务是准确推断该因果DAG,即使在存在噪声的情况下。为实现此目标,智能体执行干预,利用因果推断技术分析观测到变化的统计显著性。结果表明,除了在最有限的资源条件下,DODO在所有情况下均优于纯观测方法。DODO通常能够以低至零误差重建因果图的结构。在最具挑战性的配置中,DODO以+0.25 F1分数的优势超越了最佳基线方法。