Multi-party computing (MPC) has been gaining popularity over the past years as a secure computing model, particularly for machine learning (ML) inference. Compared with its competitors, MPC has fewer overheads than homomorphic encryption (HE) and has a more robust threat model than hardware-based trusted execution environments (TEE) such as Intel SGX. Despite its apparent advantages, MPC protocols still pay substantial performance penalties compared to plaintext when applied to ML algorithms. The overhead is due to added computation and communication costs. For multiplications that are ubiquitous in ML algorithms, MPC protocols add 32x more computational costs and 1 round of broadcasting among MPC servers. Moreover, ML computations that have trivial costs in plaintext, such as Softmax, ReLU, and other non-linear operations become very expensive due to added communication. Those added overheads make MPC less palatable to deploy in real-time ML inference frameworks, such as speech translation. In this work, we present MPC-Pipe, an MPC pipeline inference technique that uses two ML-specific approaches. 1) inter-linear-layer pipeline and 2) inner layer pipeline. Those two techniques shorten the total inference runtime for machine learning models. Our experiments have shown to reduce ML inference latency by up to 12.6% when model weights are private and 14.48\% when model weights are public, compared to current MPC protocol implementations.
翻译:多年来,多党计算(MPC)作为一种安全的计算模型越来越受欢迎,特别是机器学习(ML)推断。与其竞争者相比,多党计算(MPC)的间接费用比同质加密(HE)要少,而且比Intel SGX等基于硬件的可信任执行环境(TEE)更有活力的威胁模型。尽管具有明显的优势,但多党计算协议仍然支付与适用于ML算法的平文本相比的大幅性能罚款。管理费用是由于增加的计算和通信成本。对于在ML算法、MPC协议增加32x计算成本和在MPC服务器中进行1轮广播的倍增。此外,多党计算(MOC)比普通文本(Softmax,ReLU)和其他非线性操作的成本要低得多。尽管增加通信的好处,但多党协议仍然支付与普通文本相比高的性能罚款。 管理费用是由于增加的计算和通信成本增加。 管理费用是由于增加的计算和通信成本增加。对于在ML算法中普遍存在的倍增倍增乘量,在MPC管道服务器中将MPC管道重量增加32。