Secure multiparty computation (MPC) has been proposed to allow multiple mutually distrustful data owners to jointly train machine learning (ML) models on their combined data. However, the datasets used for training ML models might be under the control of an adversary mounting a data poisoning attack, and MPC prevents inspecting training sets to detect poisoning. We show that multiple MPC frameworks for private ML training are susceptible to backdoor and targeted poisoning attacks. To mitigate this, we propose SafeNet, a framework for building ensemble models in MPC with formal guarantees of robustness to data poisoning attacks. We extend the security definition of private ML training to account for poisoning and prove that our SafeNet design satisfies the definition. We demonstrate SafeNet's efficiency, accuracy, and resilience to poisoning on several machine learning datasets and models. For instance, SafeNet reduces backdoor attack success from 100% to 0% for a neural network model, while achieving 39x faster training and 36x less communication than the four-party MPC framework of Dalskov et al.
翻译:安全多功能计算(MPC)建议允许多个互不信任的数据所有者联合培训机器学习(ML)综合数据模型,然而,用于培训ML模型的数据集可能处于对手发动数据中毒攻击的控制之下,而MPC防止检查用于检测中毒的成套培训。我们显示,私人ML培训的多个MPC框架容易后门和有针对性的中毒袭击。为了减轻这一影响,我们提议SafeNet(SafeNet),这是在MPC建立组合模型的框架,正式保证数据中毒袭击的稳健性。我们将私人ML培训的安全定义扩大到说明中毒原因,并证明我们的安全网设计符合这一定义。我们展示了SafeNet的效率、准确性和耐受若干机器学习数据集和模型中毒的能力。例如,SafeNet将神经网络模型的后门攻击成功率从100%降至0%,同时实现比Dalskov等人的四党MPC框架更快的39x培训和36x通信减少。