Combining Federated Learning (FL) with a Trusted Execution Environment (TEE) is a promising approach for realizing privacy-preserving FL, which has garnered significant academic attention in recent years. Implementing the TEE on the server side enables each round of FL to proceed without exposing the client's gradient information to untrusted servers. This addresses usability gaps in existing secure aggregation schemes as well as utility gaps in differentially private FL. However, to address the issue using a TEE, the vulnerabilities of server-side TEEs need to be considered -- this has not been sufficiently investigated in the context of FL. The main technical contribution of this study is the analysis of the vulnerabilities of TEE in FL and the defense. First, we theoretically analyze the leakage of memory access patterns, revealing the risk of sparsified gradients, which are commonly used in FL to enhance communication efficiency and model accuracy. Second, we devise an inference attack to link memory access patterns to sensitive information in the training dataset. Finally, we propose an oblivious yet efficient aggregation algorithm to prevent memory access pattern leakage. Our experiments on real-world data demonstrate that the proposed method functions efficiently in practical scales.
翻译:将受信任执行环境(TEE)与联邦学习(FL)结合起来,是实现隐私保护FL的一种有前途的方法,在最近几年引起了广泛的学术关注。在服务器端实现TEE可以使每轮FL在不将客户端的梯度信息暴露给不受信任的服务器的情况下进行。这解决了现有安全的聚合方案的可用性差距以及差分私人FL的效用差距。然而,为了利用TEE解决此问题,需要考虑服务器端TEE的漏洞,这在FL的情况下尚未得到充分的研究。本研究的主要技术贡献是分析了FL中TEE的漏洞和防御。首先,我们从理论上分析了内存访问模式的泄漏,揭示了使用FL中常用的稀疏梯度的风险,以增强通信效率和模型准确度。其次,我们设计了一种推理攻击,将内存访问模式与训练数据集中的敏感信息联系起来。最后,我们提出了一种无意识而有效的聚合算法,以防止内存访问模式泄漏。我们对真实数据的实验表明,所提出的方法在实际规模上运作高效。