Knowledge Graph Embedding (KGE) is a fundamental technique that extracts expressive representation from knowledge graph (KG) to facilitate diverse downstream tasks. The emerging federated KGE (FKGE) collaboratively trains from distributed KGs held among clients while avoiding exchanging clients' sensitive raw KGs, which can still suffer from privacy threats as evidenced in other federated model trainings (e.g., neural networks). However, quantifying and defending against such privacy threats remain unexplored for FKGE which possesses unique properties not shared by previously studied models. In this paper, we conduct the first holistic study of the privacy threat on FKGE from both attack and defense perspectives. For the attack, we quantify the privacy threat by proposing three new inference attacks, which reveal substantial privacy risk by successfully inferring the existence of the KG triple from victim clients. For the defense, we propose DP-Flames, a novel differentially private FKGE with private selection, which offers a better privacy-utility tradeoff by exploiting the entity-binding sparse gradient property of FKGE and comes with a tight privacy accountant by incorporating the state-of-the-art private selection technique. We further propose an adaptive privacy budget allocation policy to dynamically adjust defense magnitude across the training procedure. Comprehensive evaluations demonstrate that the proposed defense can successfully mitigate the privacy threat by effectively reducing the success rate of inference attacks from $83.1\%$ to $59.4\%$ on average with only a modest utility decrease.
翻译:知识图谱嵌入(KGE)是一种从知识图谱(KG)中提取表达式表示以促进各种下游任务的基本技术。新兴的联邦KGE(FKGE)在避免交换客户端敏感原始KG的情况下从分布式客户持有的KG中集体培训,尽管在其他联邦模型训练中(例如神经网络)仍可能受到隐私威胁。然而,FKGE上的这种隐私威胁的量化和抵御仍未得到探究,因为FKGE具有先前研究的模型所没有的独特属性。在本文中,我们从攻击和防御角度开展了FKGE上的首个隐私威胁的整体研究。对于攻击,我们提出了三种新的推理攻击来量化隐私威胁,这些攻击通过成功推断受害客户端的KG三元组存在来揭示相当大的隐私风险。对于防御,我们提出了DP-Flames,一种利用FKGE实体绑定稀疏梯度属性、具有私有选择的新型差分隐私FKGE,它通过结合最先进的私有选择技术,提供更好的隐私效用权衡,并带有紧密的隐私账户。我们还提出了自适应隐私预算分配策略,以在训练过程中动态调整防御强度。全面的评估表明,所提出的防御可以通过将成功率从83.1%降低到59.4%的平均值,有效减少推理攻击的隐私威胁,只会带来适度的效用降低。