Machine learning models are increasingly used by businesses and organizations around the world to automate tasks and decision-making. Trained on potentially sensitive datasets, machine learning models have been shown to leak information about individuals in the dataset as well as global dataset information. We here take research in dataset property inference attacks one step further by proposing a new attack against ML models: a dataset correlation inference attack, where an attacker's goal is to infer the correlation between input variables of a model. We first show that an attacker can exploit the spherical parametrization of correlation matrices, to make an informed guess. This means that using only the correlation between the input variables and the target variable, an attacker can infer the correlation between two input variables much better than a random guess baseline. We propose a second attack which exploits the access to a machine learning model using shadow modeling to refine the guess. Our attack uses Gaussian copula-based generative modeling to generate synthetic datasets with a wide variety of correlations in order to train a meta-model for the correlation inference task. We evaluate our attack against Logistic Regression and Multi-layer perceptron models and show it to outperform the model-less attack. Our results show that the accuracy of the second, machine learning-based attack decreases with the number of variables and converges towards the accuracy of the model-less attack. However, correlations between input variables which are highly correlated with the target variable are more vulnerable regardless of the number of variables. Our work bridges the gap between what can be considered a global leakage about the training dataset and individual-level leakages. When coupled with marginal leakage attacks,it might also constitute a first step towards dataset reconstruction.
翻译:世界各地企业和组织越来越多地使用机器学习模型来使任务和决策自动化。 在潜在敏感的数据集上, 机器学习模型显示在数据集和全球数据集信息中泄漏个人信息。 我们在这里对数据集属性推断攻击进行更进一步的研究, 提议对 ML 模型进行新的攻击: 数据集相关性推断攻击, 攻击者的目标是推断模型输入变量之间的相互关系。 我们首先显示攻击者可以利用相关关系矩阵对相关矩阵进行球状对流的猜测, 以便做出知情的猜测。 这意味着仅使用输入变量与目标变量之间的相互关系, 攻击者可以推断两个输入变量之间的相互关系比随机猜测基线要好得多。 我们提议进行第二次攻击, 利用机器学习模型访问的机会, 利用阴影模型进行更精确的推断。 我们的攻击利用高斯的基云式模型的变相模型模型模型模型来生成合成数据集, 高度不同的对应关系, 用来对攻击目标变量变量与目标变量变量之间的关联性猜测。 我们的变异性数据变异性数据显示, 我们的变异性数据显示我们攻击的变异性数据显示我们攻击的变异性, 显示我们攻击的变异性变异性数据显示我们攻击的变异的变变。