Modern machine learning approaches excel in static settings where a large amount of i.i.d. training data are available for a given task. In a dynamic environment, though, an intelligent agent needs to be able to transfer knowledge and re-use learned components across domains. It has been argued that this may be possible through causal models, aiming to mirror the modularity of the real world in terms of independent causal mechanisms. However, the true causal structure underlying a given set of data is generally not identifiable, so it is desirable to have means to quantify differences between models (e.g., between the ground truth and an estimate), on both the observational and interventional level. In the present work, we introduce the Interventional Kullback-Leibler (IKL) divergence to quantify both structural and distributional differences between models based on a finite set of multi-environment distributions generated by interventions from the ground truth. Since we generally cannot quantify all differences between causal models for every finite set of interventional distributions, we propose a sufficient condition on the intervention targets to identify subsets of observed variables on which the models provably agree or disagree.
翻译:现代机器学习方法在静态环境中非常出色,在静态环境中,为某项任务提供了大量的培训数据。但在动态环境中,智能剂需要能够转让知识和再利用跨领域的学习组件。有人争辩说,这可以通过因果模型来实现,目的是用独立的因果机制来反映真实世界的模块性。然而,某组数据背后的真正因果结构一般无法识别,因此最好有办法量化各种模型(例如地面真相和估计数字)在观测和干预层面的差异。在目前的工作中,我们引入了干预性库尔背利贝尔(IKL)差异,以量化基于从地面真相干预中产生的一套有限的多环境分布模型的结构性和分布差异。由于我们一般无法量化每一组有限的干预分布的因果模型之间的所有差异,因此,我们建议对干预目标提出一个充分的条件,以确定观察到的变量的子集,这些变量是模型所同意或不同意的。