Cloud-edge collaborative inference approach splits deep neural networks (DNNs) into two parts that run collaboratively on resource-constrained edge devices and cloud servers, aiming at minimizing inference latency and protecting data privacy. However, even if the raw input data from edge devices is not directly exposed to the cloud, state-of-the-art attacks targeting collaborative inference are still able to reconstruct the raw private data from the intermediate outputs of the exposed local models, introducing serious privacy risks. In this paper, a secure privacy inference framework for cloud-edge collaboration is proposed, termed CIS, which supports adaptively partitioning the network according to the dynamically changing network bandwidth and fully releases the computational power of edge devices. To mitigate the influence introduced by private perturbation, CIS provides a way to achieve differential privacy protection by adding refined noise to the intermediate layer feature maps offloaded to the cloud. Meanwhile, with a given total privacy budget, the budget is reasonably allocated by the size of the feature graph rank generated by different convolution filters, which makes the inference in the cloud robust to the perturbed data, thus effectively trade-off the conflicting problem between privacy and availability. Finally, we construct a real cloud-edge collaborative inference computing scenario to verify the effectiveness of inference latency and model partitioning on resource-constrained edge devices. Furthermore, the state-of-the-art cloud-edge collaborative reconstruction attack is used to evaluate the practical availability of the end-to-end privacy protection mechanism provided by CIS.
翻译:然而,即使边缘装置的原始输入数据没有直接暴露在云层中,但针对协作推断的最先进的攻击仍然能够从暴露的当地模型的中间产出中重建原始私人数据,从而带来严重的隐私风险。本文提议建立一个由资源限制的边缘装置和云端服务器共同运行的安全隐私推断框架,目的是根据动态变化的网络带宽支持对网络进行适应性分割,并充分释放边缘装置的计算能力。然而,即使边缘装置的原始输入数据没有直接暴露在云层中,但针对协作推断的最先进的攻击仍然能够将原始私人数据从暴露的当地模型的中间产出中重建出来,从而带来严重的隐私风险。在本文中,由不同变迁过滤器生成的地貌等级的大小合理分配了预算,这使得云层对透视层带动态带宽宽宽广,从而有效地释放了边缘装置的计算能力。