Mediation analysis learns the causal effect transmitted via mediator variables between treatments and outcomes and receives increasing attention in various scientific domains to elucidate causal relations. Most existing works focus on point-exposure studies where each subject only receives one treatment at a single time point. However, there are a number of applications (e.g., mobile health) where the treatments are sequentially assigned over time and the dynamic mediation effects are of primary interest. Proposing a reinforcement learning (RL) framework, we are the first to evaluate dynamic mediation effects in settings with infinite horizons. We decompose the average treatment effect into an immediate direct effect, an immediate mediation effect, a delayed direct effect, and a delayed mediation effect. Upon the identification of each effect component, we further develop robust and semi-parametrically efficient estimators under the RL framework to infer these causal effects. The superior performance of the proposed method is demonstrated through extensive numerical studies, theoretical results, and an analysis of a mobile health dataset.
翻译:调解分析通过调解者变量了解治疗和结果之间的因果影响,并在各种科学领域日益受到重视,以阐明因果关系。大多数现有工作侧重于点暴露研究,每个对象在某一时间点只接受一种治疗。然而,有一些应用(如流动健康),治疗是按顺序排列的,动态调解效应是首要利益所在。提出强化学习框架,我们首先评估无限视野环境中动态调解效应。我们将平均治疗效应分解为直接的立即效果、即时调解效应、延迟直接效应和延迟调解效应。在确定每种效果组成部分后,我们在RL框架下进一步开发稳健和半对称高效的估算器,以推断这些因果关系。通过广泛的数字研究、理论结果和对流动健康数据集的分析,可以证明拟议方法的优异性表现。