When neural network model and data are outsourced to cloud server for inference, it is desired to preserve the confidentiality of model and data as the involved parties (i.e., cloud server, model providing client and data providing client) may not trust mutually. Solutions were proposed based on multi-party computation, trusted execution environment (TEE) and leveled or fully homomorphic encryption (LHE/FHE), but their limitations hamper practical application. We propose a new framework based on synergistic integration of LHE and TEE, which enables collaboration among mutually-untrusted three parties, while minimizing the involvement of (relatively) resource-constrained TEE and allowing the full utilization of the untrusted but more resource-rich part of server. We also propose a generic and efficient LHE-based inference scheme as an important performance-determining component of the framework. We implemented/evaluated the proposed system on a moderate platform and show that, our proposed scheme is more applicable/scalable to various settings, and has better performance, compared to the state-of-the-art LHE-based solutions.
翻译:当将神经网络模型和数据外包给云服务器以便作出推断时,希望保持模型和数据的保密性,因为有关各方(即云服务器、提供客户和数据提供客户的模型)可能互不信任;提出了基于多方计算、可信赖的执行环境(TEE)以及平分或完全同质加密(LHE/FHE)的解决办法,但其局限性妨碍了实际应用;我们提议了一个基于LHE和TEE协同整合的新框架,使互不信任的三方之间能够合作,同时尽量减少(相对而言)受资源限制的TEE的参与,并允许充分利用服务器中不受信任但资源更丰富的部分;我们还提议了一个通用和高效的LHE推论办法,作为框架一个重要的绩效确定组成部分;我们在一个中度平台上实施/评价了拟议的系统,并表明,我们提议的计划更适用于/可适用于各种环境,而且与基于最新水平的LHE的解决方案相比,其性能更好。