This paper investigates the problem of forecasting multivariate aggregated human mobility while preserving the privacy of the individuals concerned. Differential privacy, a state-of-the-art formal notion, has been used as the privacy guarantee in two different and independent steps when training deep learning models. On one hand, we considered \textit{gradient perturbation}, which uses the differentially private stochastic gradient descent algorithm to guarantee the privacy of each time series sample in the learning stage. On the other hand, we considered \textit{input perturbation}, which adds differential privacy guarantees in each sample of the series before applying any learning. We compared four state-of-the-art recurrent neural networks: Long Short-Term Memory, Gated Recurrent Unit, and their Bidirectional architectures, i.e., Bidirectional-LSTM and Bidirectional-GRU. Extensive experiments were conducted with a real-world multivariate mobility dataset, which we published openly along with this paper. As shown in the results, differentially private deep learning models trained under gradient or input perturbation achieve nearly the same performance as non-private deep learning models, with loss in performance varying between $0.57\%$ to $2.8\%$. The contribution of this paper is significant for those involved in urban planning and decision-making, providing a solution to the human mobility multivariate forecast problem through differentially private deep learning models.
翻译:本文调查了在保护相关个人隐私的同时预测多变的人类流动总量的问题。 不同隐私是一种最先进的正式概念,在培训深层学习模式时,在两种不同的独立步骤中使用了不同隐私作为隐私保障。 一方面,我们考虑了\ textit{ 快速扰动},它使用差异式私人随机梯度梯度下限算法来保障学习阶段每个时间序列样本的隐私。 另一方面,我们考虑了在应用任何学习之前,在每组样本中增加差异性隐私保障。我们比较了四种最先进的常规神经网络:长期短期记忆、Gated 常规单元及其双向结构,即双向性偏向-LSTM和双向-GRU。 我们用一个真实的多变式流动数据集进行了广泛的实验,我们公开公布了这些数据。 如结果所示,在梯度或投入下培训的不同私人深层隐私学习模式,在深度或深度投入中, 几乎实现了这种不明显的城市学习模式的绩效, 也就是通过深度的深度学习模式,在深度投资中, 20美元 。