The Shapley value (SV) is a fair and principled metric for contribution evaluation in cross-silo federated learning (cross-silo FL), in which organizations, i.e., clients, collaboratively train prediction models with the coordination of a parameter server. However, existing SV calculation methods for FL assume that the server can access the raw FL models and public test data, which might not be a valid assumption in practice given the emerging privacy attacks on FL models and that test data might be clients' private assets. Hence, in this paper, we investigate the problem of secure SV calculation for cross-silo FL. We first propose a one-server solution, HESV, which is based solely on homomorphic encryption (HE) for privacy protection and has some considerable limitations in efficiency. To overcome these limitations, we further propose an efficient two-server protocol, SecSV, which has the following novel features. First, SecSV utilizes a hybrid privacy protection scheme to avoid ciphertext-ciphertext multiplications between test data and models, which are extremely expensive under HE. Second, a more efficient secure matrix multiplication method is proposed for SecSV. Third, SecSV strategically identifies and skips some test samples without significantly affecting the evaluation accuracy. Our experiments demonstrate that SecSV is 5.4-25.7 times as fast as HESV while sacrificing a limited loss in the accuracy of calculated SVs.
翻译:沙普利值(SV)是跨筒联结学习(跨筒联校 FL)的缴款评价的一个公平而有原则的衡量标准,各组织,即客户,在跨筒联学习(跨筒联校 FL)中合作培训预测模型,并协调参数服务器;然而,现有的FLSV计算方法假定服务器可以访问原始FL模型和公共测试数据,鉴于对FL模型的隐私攻击正在出现,测试数据可能是客户的私人资产,这在实践中可能不是一种有效的假设。因此,我们在本文件中调查了跨筒联联学习(跨筒联校FL)的安全SV计算问题。我们首先提出了一个只基于同质加密(HE)的单一服务器解决方案,用于保护隐私,在效率方面也存在一些相当大的限制。为了克服这些限制,我们进一步提议一个高效的双服务器协议(SecSV),它具有以下新特点。SecV使用一种混合隐私保护计划,以避免跨筒V数据和模型之间出现加密的倍增。我们提出的测试数据在HEV的精确性试验中是极其昂贵的。