Backdoor attacks are a major concern in federated learning (FL) pipelines where training data is sourced from untrusted clients over long periods of time (i.e., continual learning). Preventing such attacks is difficult because defenders in FL do not have access to raw training data. Moreover, in a phenomenon we call backdoor leakage, models trained continuously eventually suffer from backdoors due to cumulative errors in backdoor defense mechanisms. We propose a novel framework for defending against backdoor attacks in the federated continual learning setting. Our framework trains two models in parallel: a backbone model and a shadow model. The backbone is trained without any defense mechanism to obtain good performance on the main task. The shadow model combines recent ideas from robust covariance estimation-based filters with early-stopping to control the attack success rate even as the data distribution changes. We provide theoretical motivation for this design and show experimentally that our framework significantly improves upon existing defenses against backdoor attacks.
翻译:后门攻击是联邦学习(FL)管道中的一个主要问题,在后门攻击中,培训数据来自长期得不到信任的客户(即持续学习)。防止这种攻击是困难的,因为FL的维护者无法获得原始培训数据。此外,在我们称之为后门泄漏的现象中,受过训练的模型最终会因为后门防御机制的累积错误而成为后门攻击的对象。我们提出了一个在联邦不断学习的环境下防御后门攻击的新框架。我们的框架平行地培训了两个模型:一个主干模型和一个影子模型。骨干在没有任何防御机制的情况下接受了培训,以取得主要任务的良好表现。影子模型将来自强有力的共变估计过滤器的最新想法与早期停止控制攻击成功率相结合,即使数据分配的变化也是如此。我们为这一设计提供了理论动机,并实验性地表明我们的框架大大改进了现有的对后门攻击的防御。