This paper tackles the problem of ensuring training data privacy in a federated learning context. Relying on Fully Homomorphic Encryption (FHE) and Differential Privacy (DP), we propose a secure framework addressing an extended threat model with respect to privacy of the training data. Notably, the proposed framework protects the privacy of the training data from all participants, namely the training data owners and an aggregating server. In details, while homomorphic encryption blinds a semi-honest server at learning stage, differential privacy protects the data from semi-honest clients participating in the training process as well as curious end-users with black-box or white-box access to the trained model. This paper provides with new theoretical and practical results to enable these techniques to be effectively combined. In particular, by means of a novel stochastic quantization operator, we prove differential privacy guarantees in a context where the noise is quantified and bounded due to the use of homomorphic encryption. The paper is concluded by experiments which show the practicality of the entire framework in spite of these interferences in terms of both model quality (impacted by DP) and computational overheads (impacted by FHE).
翻译:本文探讨了在联合学习背景下确保培训数据隐私的问题。我们依据全单态加密(FHE)和差异隐私(DP)提出一个安全框架,解决培训数据隐私方面的长期威胁模式,特别是,拟议框架保护培训数据来自所有参与者,即培训数据所有人和汇总服务器的隐私。具体而言,同单式加密在学习阶段使半诚实服务器失灵,而不同的隐私保护了参加培训进程的半诚实客户以及黑盒或白盒访问培训模式的好奇最终用户的数据。本文提供了新的理论和实际结果,使这些技术能够有效结合。特别是,在噪音被量化并因使用同质加密而受约束的情况下,我们通过创新的静态四分化操作器,证明隐私保障存在差异。文件的结尾是实验,从模型质量(受DP影响)和计算间接费用(受FHE影响)的角度来看,尽管存在这些干扰,但整个框架仍具有实用性。