Preserving the individuals' privacy in sharing spatial-temporal datasets is critical to prevent re-identification attacks based on unique trajectories. Existing privacy techniques tend to propose ideal privacy-utility tradeoffs, however, largely ignore the fairness implications of mobility models and whether such techniques perform equally for different groups of users. The quantification between fairness and privacy-aware models is still unclear and there barely exists any defined sets of metrics for measuring fairness in the spatial-temporal context. In this work, we define a set of fairness metrics designed explicitly for human mobility, based on structural similarity and entropy of the trajectories. Under these definitions, we examine the fairness of two state-of-the-art privacy-preserving models that rely on GAN and representation learning to reduce the re-identification rate of users for data sharing. Our results show that while both models guarantee group fairness in terms of demographic parity, they violate individual fairness criteria, indicating that users with highly similar trajectories receive disparate privacy gain. We conclude that the tension between the re-identification task and individual fairness needs to be considered for future spatial-temporal data analysis and modelling to achieve a privacy-preserving fairness-aware setting.
翻译:在共享时空数据中保护个人隐私对于防止基于唯一轨迹的再识别攻击至关重要。现有的隐私技术往往提出理想的隐私-效用折衷方案,但大大忽略了移动模型的公平性影响,以及这种技术是否对不同用户群体同样有效。空间-时间背景下公平性与隐私感知模型之间的量化关系仍然不清楚,几乎没有任何定义明确的用于衡量公平性的指标集。在这项工作中,我们针对人类移动性定义了一组公平度量标准,基于轨迹的结构相似性和熵。在这些定义下,我们考察了两个最先进的依赖于GAN和表示学习来减少用户再识别率的隐私保护模型的公平性。我们的研究结果表明,虽然这两个模型都保证了公民群体之间的公平性,但它们违反了个人公平标准,表明具有高度相似轨迹的用户收到了不同的隐私保护收益。我们得出结论,隐私保护和公平性之间的紧张关系需要在未来的空间-时间数据分析和建模中得到考虑,以实现隐私保护公平意识环境。