Federated Learning (FL) is a scheme for collaboratively training Deep Neural Networks (DNNs) with multiple data sources from different clients. Instead of sharing the data, each client trains the model locally, resulting in improved privacy. However, recently so-called targeted poisoning attacks have been proposed that allow individual clients to inject a backdoor into the trained model. Existing defenses against these backdoor attacks either rely on techniques like Differential Privacy to mitigate the backdoor, or analyze the weights of the individual models and apply outlier detection methods that restricts these defenses to certain data distributions. However, adding noise to the models' parameters or excluding benign outliers might also reduce the accuracy of the collaboratively trained model. Additionally, allowing the server to inspect the clients' models creates a privacy risk due to existing knowledge extraction methods. We propose CrowdGuard, a model filtering defense, that mitigates backdoor attacks by leveraging the clients' data to analyze the individual models before the aggregation. To prevent data leaks, the server sends the individual models to secure enclaves, running in client-located Trusted Execution Environments. To effectively distinguish benign and poisoned models, even if the data of different clients are not independently and identically distributed (non-IID), we introduce a novel metric called HLBIM to analyze the outputs of the DNN's hidden layers. We show that the applied significance-based detection algorithm combined can effectively detect poisoned models, even in non-IID scenarios. We show in our extensive evaluation that CrowdGuard can effectively mitigate targeted poisoning attacks and achieve in various scenarios a True-Positive-Rate of 100% and a True-Negative-Rate of 100%.
翻译:联邦学习联合会(FL) 是一个协作培训深神经网络(DNNS) 的计划, 由来自不同客户的多个数据源共同培训 深神经网络(DNNS) 。 然而, 最近提出了所谓的定点中毒袭击, 允许个别客户将后门注入经过培训的模型。 针对这些后门袭击的现有防御要么依靠“差异隐私” 等技术来缓解后门袭击, 要么分析个体模型的重量, 并应用外部检测方法, 将这些防御限制在某些数据分布上。 但是, 在模型参数中添加噪音或排除良异端外端器, 反而会有效降低经过协作训练的模式的准确性。 此外, 允许服务器检查客户的模型会因现有的知识提取方法而造成隐私风险。 我们提议CrowdGuardGuard, 通过利用客户的数据来分析基于总和之前的单个模型, 来减轻后门攻击。 为了防止数据泄漏, 服务器让个人模型实现飞地, 在客户定位的信任执行环境中运行, 排除良性外外外外外外外外外外外外外外外外外值。, 我们方的直方的直径二号的内测算数据, 。