Gaussian processes (GPs) are non-parametric Bayesian models that are widely used for diverse prediction tasks. Previous work in adding strong privacy protection to GPs via differential privacy (DP) has been limited to protecting only the privacy of the prediction targets (model outputs) but not inputs. We break this limitation by introducing GPs with DP protection for both model inputs and outputs. We achieve this by using sparse GP methodology and publishing a private variational approximation on known inducing points. The approximation covariance is adjusted to approximately account for the added uncertainty from DP noise. The approximation can be used to compute arbitrary predictions using standard sparse GP techniques. We propose a method for hyperparameter learning using a private selection protocol applied to validation set log-likelihood. Our experiments demonstrate that given sufficient amount of data, the method can produce accurate models under strong privacy protection.
翻译:Gaussian 过程(GPs) 是非参数性的贝耶斯模型,广泛用于不同的预测任务。以前通过不同的隐私(DP)为GP增加强力的隐私保护的工作仅限于保护预测目标的隐私(模型产出),而不是保护投入。我们通过对模型投入和产出采用DP保护来打破这一限制。我们通过使用稀疏的GP方法并在已知诱导点上公布一种私人的变异近似来实现这一目标。近似差数被调整为大致考虑到DP噪音增加的不确定性。近似差可用于使用标准的稀有GP技术计算任意预测。我们提出使用用于验证日志相似性的私人选择协议进行超光谱学习的方法。我们的实验表明,如果数据足够多,该方法可以在强大的隐私保护下产生准确模型。