Oblivious inference is the task of outsourcing a ML model, like neural-networks, without disclosing critical and sensitive information, like the model's parameters. One of the most prominent solutions for secure oblivious inference is based on a powerful cryptographic tools, like Homomorphic Encryption (HE) and/or multi-party computation (MPC). Even though the implementation of oblivious inference systems schemes has impressively improved the last decade, there are still significant limitations on the ML models that they can practically implement. Especially when both the ML model and the input data's confidentiality must be protected. In this paper, we introduce the notion of partially oblivious inference. We empirically show that for neural network models, like CNNs, some information leakage can be acceptable. We therefore propose a novel trade-off between security and efficiency. In our research, we investigate the impact on security and inference runtime performance from the CNN model's weights partial leakage. We experimentally demonstrate that in a CIFAR-10 network we can leak up to $80\%$ of the model's weights with practically no security impact, while the necessary HE-mutliplications are performed four times faster.
翻译:显而易见的推论是外包ML模型的任务,如神经网络,而不披露关键和敏感信息,就像模型的参数一样。安全隐蔽推断的最突出解决办法之一是基于强大的加密工具,如单式加密和/或多方计算。尽管在过去十年中,执行隐含的推断系统计划已经显著改善了,但它们实际可以实际执行的ML模型仍然有相当大的局限性。特别是当ML模型和输入数据的保密性必须受到保护时。在本文中,我们引入了部分隐蔽推断的概念。我们从经验上表明,对于神经网络模型,如CNNs,某些信息渗漏是可以接受的。因此,我们提出了安全和效率之间的新交易。在我们的研究中,我们研究了CNN模型的重量部分渗漏对安全和推断的运行性能的影响。我们实验性地证明,在CIFAR-10网络中,我们可以将模型的重量泄漏到80美分之80美分,而安全性影响却在实际的四倍增速度上。