Differential privacy (DP) provides a robust model to achieve privacy guarantees for released information. We examine the protection potency of sanitized multi-dimensional frequency distributions via DP randomization mechanisms against homogeneity attack (HA). HA allows adversaries to obtain the exact values on sensitive attributes for their targets without having to identify them from the released data. We propose measures for disclosure risk from HA and derive closed-form relationships between the privacy loss parameters in DP and the disclosure risk from HA. The availability of the closed-form relationships assists understanding the abstract concepts of DP and privacy loss parameters by putting them in the context of a concrete privacy attack and offers a perspective for choosing privacy loss parameters when employing DP mechanisms in information sanitization and release in practice. We apply the closed-form mathematical relationships in real-life datasets to demonstrate the assessment of disclosure risk due to HA on differentially private sanitized frequency distributions at various privacy loss parameters.
翻译:不同隐私(DP)提供了一个强有力的模式,以实现对发布信息的隐私保障。我们研究了通过DP随机随机处理机制保护被清洁的多维频率分布防止同源性攻击(HA)的有效性。HA允许对手获得目标敏感属性的确切值,而不必从发布的数据中识别这些特性。我们建议了由HA披露风险的措施,并从DP的隐私损失参数与HA披露风险之间得出封闭式关系。闭式关系的存在有助于理解DP的抽象概念和隐私损失参数,将其置于具体的隐私攻击中,并提供了一个在实际使用DP机制进行信息清洁和发布时选择隐私损失参数的视角。我们在现实生活数据集中应用封闭式数学关系,以展示对HA在不同隐私损失参数上差异化的私人保密频率分布对披露风险的评估。