The privacy concerns of providing deep learning inference as a service have underscored the need for private inference (PI) protocols that protect users' data and the service provider's model using cryptographic methods. Recently proposed PI protocols have achieved significant reductions in PI latency by moving the computationally heavy homomorphic encryption (HE) parts to an offline/pre-compute phase. Paired with recent optimizations that tailor networks for PI, these protocols have achieved performance levels that are tantalizingly close to being practical. In this paper, we conduct a rigorous end-to-end characterization of PI protocols and optimization techniques and find that the current understanding of PI performance is overly optimistic. Specifically, we find that offline storage costs of garbled circuits (GC), a key cryptographic protocol used in PI, on user/client devices are prohibitively high and force much of the expensive offline HE computation to the online phase, resulting in a 10-1000$\times$ increase to PI latency. We propose a modified PI protocol that significantly reduces client-side storage costs for a small increase in online latency. Evaluated end-to-end, the modified protocol outperforms current protocols by reducing the mean PI latency by $4\times$ for ResNet18 on TinyImageNet. We conclude with a discussion of several recently proposed PI optimizations in light of the findings and note many actually increase PI latency when evaluated from an end-to-end perspective.
翻译:提供深层学习推断作为服务的隐私问题突出表明,需要使用加密方法保护用户数据和服务提供者模型的私人从端到端程序(PI)说明,最近提出的PI协议通过将计算中重同质加密部件(HE)的离线/预计算阶段,大大降低了PI的延缩度;最近优化了为PI定制网络,这些协议达到了接近实际的性能水平;在本文件中,我们从严格的端到端描述PI协议和优化技术,发现目前对PI绩效的理解过于乐观;具体地说,我们发现,由于将计算中重的重同质加密部件(He)部分转移到离线/预计算阶段,在用户/客户设备上使用的关键加密协议(He)的离线存储成本太高,而且将昂贵的离线网络网络网络网络网络网络的计算大部分都强制到在线阶段,导致PILE的升至10-1000美元时间。我们建议修改PI协议,以大幅降低当前客户端讨论成本,通过对在线最后协议进行小规模评估,从而降低目前对协议进行小规模升级的升级。