Federated learning has emerged as a popular paradigm for collaboratively training a model from data distributed among a set of clients. This learning setting presents, among others, two unique challenges: how to protect privacy of the clients' data during training, and how to ensure integrity of the trained model. We propose a two-pronged solution that aims to address both challenges under a single framework. First, we propose to create secure enclaves using a trusted execution environment (TEE) within the server. Each client can then encrypt their gradients and send them to verifiable enclaves. The gradients are decrypted within the enclave without the fear of privacy breaches. However, robustness check computations in a TEE are computationally prohibitive. Hence, in the second step, we perform a novel gradient encoding that enables TEEs to encode the gradients and then offloading Byzantine check computations to accelerators such as GPUs. Our proposed approach provides theoretical bounds on information leakage and offers a significant speed-up over the baseline in empirical evaluation.
翻译:联邦学习已成为合作培训模式的流行范例,这种模式来自一组客户之间分配的数据。这种学习环境除其他外,提出了两个独特的挑战:如何在培训期间保护客户数据的隐私,以及如何确保培训模式的完整性。我们提出了一个双管齐下的解决办法,目的是在单一框架内应对这两项挑战。首先,我们提议利用服务器内一个可信赖的执行环境(TEE)来创建安全的飞地。每个客户然后可以加密其梯度并将其发送到可核查的飞地。梯度在飞地内解密,而不必担心隐私被侵犯。然而,在TEE中进行稳健性检查的计算是计算上令人窒息的。因此,在第二步,我们执行一个新的梯度编码,使TEE能够对梯度进行编码,然后将Byzantine校验计算从GPUs等加速器上卸载。我们提出的方法提供了信息泄漏的理论界限,并提供了在经验评估基线上的重大速度。