Local differential privacy (LDP) is a differential privacy (DP) paradigm in which individuals first apply a DP mechanism to their data (often by adding noise) before transmiting the result to a curator. LDP ensures strong user privacy protection because the curator does not have access to any of the user's original information. On the curator's side, however, the noise for privacy results in additional bias and variance in their analyses; thus it is of great importance for analysts to incorporate the privacy noise into valid statistical inference. In this article, we develop methodologies to infer causal effects from privatized data under the Rubin Causal Model framework. First, we present asymptotically unbiased and consistent estimators with their variance estimators and plug-in confidence intervals. Second, we develop a Bayesian nonparametric methodology along with a blocked Gibbs sampling algorithm, which performs well in terms of MSE for tight privacy budgets. Finally, we present simulation studies to evaluate the performance of our proposed frequentist and Bayesian methodologies for various privacy budgets, resulting in useful suggestions for performing causal inference for privatized data.
翻译:当地差异隐私(LDP)是一种差异性隐私模式(DP),在这种模式中,个人首先对数据适用DP机制(通常是通过添加噪音),然后才将结果传递给保管人。LDP确保强有力的用户隐私保护,因为保管人无法获得任何用户原始信息。然而,在保管人方面,隐私噪音导致进一步偏差和差异,因此分析员将隐私噪音纳入有效的统计推论非常重要。在本条中,我们制定了方法,从Rubin Causal模型框架下的私有化数据中推断因果关系。首先,我们提出了带有偏向性和一致性的估算员及其差异估计和插入信任间隔。第二,我们开发了一种巴伊斯非参数,以及一种封闭的粗略抽样算法,该算法在隐私预算方面表现良好。最后,我们提出模拟研究,以评价我们为各种隐私预算提议的常客和Bayesian方法的绩效,从而提出有用的建议,对私有化数据进行因果关系推断。