Secure multiparty computation (MPC) has been proposed to allow multiple mutually distrustful data owners to jointly train machine learning (ML) models on their combined data. However, by design, MPC protocols faithfully compute the training functionality, which the adversarial ML community has shown to leak private information and can be tampered with in poisoning attacks. In this work, we argue that model ensembles, implemented in our framework called SafeNet, are a highly MPC-amenable way to avoid many adversarial ML attacks. The natural partitioning of data amongst owners in MPC training allows this approach to be highly scalable at training time, provide provable protection from poisoning attacks, and provably defense against a number of privacy attacks. We demonstrate SafeNet's efficiency, accuracy, and resilience to poisoning on several machine learning datasets and models trained in end-to-end and transfer learning scenarios. For instance, SafeNet reduces backdoor attack success significantly, while achieving $39\times$ faster training and $36 \times$ less communication than the four-party MPC framework of Dalskov et al. Our experiments show that ensembling retains these benefits even in many non-iid settings. The simplicity, cheap setup, and robustness properties of ensembling make it a strong first choice for training ML models privately in MPC.
翻译:安全多党计算(MPC)建议允许多个互不信任的数据拥有者联合培训机器学习(ML)综合数据模型,然而,通过设计,MPC协议忠实地计算培训功能,敌对的ML社区显示,这种功能会泄露私人信息,在中毒袭击中可以被篡改。在这项工作中,我们认为,在我们称为SafeNet的框架内实施的模型组合是避免许多对抗性ML攻击的高度MPC优劣方式。在MPC培训的所有者之间进行数据自然分割,使得这一方法在培训时高度可扩展,提供可调适的防中毒袭击的保护,并可以对一些隐私攻击进行可辨别的保护。我们展示了SafeNet的效率、准确性和耐受中毒影响的能力,在终端到终端和传输学习情景中受过培训的若干机器学习数据集和模型。例如,安全网大大降低了后门攻击的成功率,同时实现了39美元快速的培训,减少了36美元通信费,比Dalskov et al的四党MPC框架低了可伸缩性,提供了可调的防范性保护,并且可以可辨防一些隐私攻击。 我们的实验显示,在最廉价的MC的简化的模型中保持了这些安全特性。