Nowadays, machine learning models and applications have become increasingly pervasive. With this rapid increase in the development and employment of machine learning models, a concern regarding privacy has risen. Thus, there is a legitimate need to protect the data from leaking and from any attacks. One of the strongest and most prevalent privacy models that can be used to protect machine learning models from any attacks and vulnerabilities is differential privacy (DP). DP is strict and rigid definition of privacy, where it can guarantee that an adversary is not capable to reliably predict if a specific participant is included in the dataset or not. It works by injecting a noise to the data whether to the inputs, the outputs, the ground truth labels, the objective functions, or even to the gradients to alleviate the privacy issue and protect the data. To this end, this survey paper presents different differentially private machine learning algorithms categorized into two main categories (traditional machine learning models vs. deep learning models). Moreover, future research directions for differential privacy with machine learning algorithms are outlined.
翻译:目前,机器学习模式和应用越来越普遍。随着机器学习模式的开发和使用迅速增加,对隐私的关注也有所上升。因此,保护数据不受泄漏和任何攻击的合理需要。保护机器学习模式免遭任何攻击和脆弱性的最强和最普遍的隐私模式之一是不同的隐私(DP)。DP是严格和僵硬的隐私定义,它能够保证对手无法可靠地预测特定参与者是否包括在数据集中。此外,它还能为数据注入噪音,无论是投入、产出、地面真相标签、客观功能,还是梯度,以缓解隐私问题和保护数据。为此,这份调查文件提出了不同的私人机器学习算法,分为两大类(传统机器学习模型与深层学习模型)。此外,还概述了与机器学习算法不同的隐私的未来研究方向。