The growing popularity of Machine Learning (ML) has led to its deployment in various sensitive domains, which has resulted in significant research focused on ML security and privacy. However, in some applications, such as autonomous driving, integrity verification of the outsourced ML workload is more critical-a facet that has not received much attention. Existing solutions, such as multi-party computation and proof-based systems, impose significant computation overhead, which makes them unfit for real-time applications. We propose Fides, a novel framework for real-time validation of outsourced ML workloads. Fides features a novel and efficient distillation technique-Greedy Distillation Transfer Learning-that dynamically distills and fine-tunes a space and compute-efficient verification model for verifying the corresponding service model while running inside a trusted execution environment. Fides features a client-side attack detection model that uses statistical analysis and divergence measurements to identify, with a high likelihood, if the service model is under attack. Fides also offers a re-classification functionality that predicts the original class whenever an attack is identified. We devised a generative adversarial network framework for training the attack detection and re-classification models. The extensive evaluation shows that Fides achieves an accuracy of up to 98% for attack detection and 94% for re-classification.
翻译:随着机器学习(ML)的普及,其在各种敏感领域中的部署已引起了对ML安全和隐私的重要研究。然而,对于自主驾驶等应用程序,外包ML工作负载的完整性验证更为关键,这一方面并没有得到太多关注。现有解决方案,例如多方计算和基于证明的系统,导致了显着的计算开销,使它们不适用于实时应用。本文提出了Fides,一种新颖的实时外包ML工作负载验证框架。Fides具有一种新颖且有效的蒸馏技术-贪心蒸馏迁移学习,用于动态蒸馏和微调验证模型,以实现相应的服务模型的验证,并在可信执行环境内运行。Fides具有一种客户端攻击检测模型,该模型使用统计分析和发散度量来识别服务模型是否被攻击的可能性很高。Fides还提供了重新分类功能,可在识别到攻击时预测原始类。我们设计了一个生成对抗网络框架来训练检测攻击和重新分类模型。广泛的评估表明,Fides实现了高达98%的攻击检测准确性和94%的重新分类准确性。