Machine learning models are known to memorize the unique properties of individual data points in a training set. This memorization capability can be exploited by several types of attacks to infer information about the training data, most notably, membership inference attacks. In this paper, we propose an approach based on information leakage for guaranteeing membership privacy. Specifically, we propose to use a conditional form of the notion of maximal leakage to quantify the information leaking about individual data entries in a dataset, i.e., the entrywise information leakage. We apply our privacy analysis to the Private Aggregation of Teacher Ensembles (PATE) framework for privacy-preserving classification of sensitive data and prove that the entrywise information leakage of its aggregation mechanism is Schur-concave when the injected noise has a log-concave probability density. The Schur-concavity of this leakage implies that increased consensus among teachers in labeling a query reduces its associated privacy cost. Finally, we derive upper bounds on the entrywise information leakage when the aggregation mechanism uses Laplace distributed noise.
翻译:机体学习模型可以记住培训集中个别数据点的独特特性。 这种记忆能力可以通过几种类型的攻击加以利用,以推断有关培训数据的信息,最明显的是成员推论攻击。在本文中,我们提议基于信息泄漏的方法,以保障会员隐私。具体地说,我们提议使用最大渗漏概念的有条件形式,以量化数据集中个别数据条目泄漏的信息,即入门信息渗漏。我们将我们的隐私分析应用到对敏感数据进行隐私保密分类的教师集体(PATE)私营框架(PATE)中,并证明当注入的噪音具有日志一致概率密度时,其聚合机制的入门信息渗漏是Schur-concave。这种渗漏的奇尔-Concave意味着教师之间在标注查询时增加共识,会降低相关的隐私费用。最后,我们从聚合机制使用拉比特号传播噪音时,我们获取入源信息渗漏的上限。