The widespread deployment of machine learning (ML) is raising serious concerns on protecting the privacy of users who contributed to the collection of training data. Differential privacy (DP) is rapidly gaining momentum in the industry as a practical standard for privacy protection. Despite DP's importance, however, little has been explored within the computer systems community regarding the implication of this emerging ML algorithm on system designs. In this work, we conduct a detailed workload characterization on a state-of-the-art differentially private ML training algorithm named DP-SGD. We uncover several unique properties of DP-SGD (e.g., its high memory capacity and computation requirements vs. non-private ML), root-causing its key bottlenecks. Based on our analysis, we propose an accelerator for differentially private ML named DiVa, which provides a significant improvement in compute utilization, leading to 2.6x higher energy-efficiency vs. conventional systolic arrays.
翻译:广泛运用机器学习(ML)正引起人们严重关切保护那些为收集培训数据作出贡献的用户的隐私问题,不同隐私(DP)作为保护隐私的实用标准,在行业中正在迅速获得势头。然而,尽管DP的重要性不大,在计算机系统社区中,对于这种新兴的ML算法对系统设计的影响,没有进行多少探讨。在这项工作中,我们对称为DP-SGD的最先进的、有差别的私人ML培训算法进行了详细的工作量定性。我们发现了DP-SGD(例如它的高记忆能力和计算要求相对于非私营ML)的几种独特特性,即:其高记忆能力和计算要求相对于非私营ML,其关键瓶颈的根部。我们根据我们的分析,建议为称为DiVa的差别私人ML加速器,该加速器在计算利用方面大有改进,导致2.6x更高的能源效率相对于常规的Systec阵列。