Data is the key factor to drive the development of machine learning (ML) during the past decade. However, high-quality data, in particular labeled data, is often hard and expensive to collect. To leverage large-scale unlabeled data, self-supervised learning, represented by contrastive learning, is introduced. The objective of contrastive learning is to map different views derived from a training sample (e.g., through data augmentation) closer in their representation space, while different views derived from different samples more distant. In this way, a contrastive model learns to generate informative representations for data samples, which are then used to perform downstream ML tasks. Recent research has shown that machine learning models are vulnerable to various privacy attacks. However, most of the current efforts concentrate on models trained with supervised learning. Meanwhile, data samples' informative representations learned with contrastive learning may cause severe privacy risks as well. In this paper, we perform the first privacy analysis of contrastive learning through the lens of membership inference and attribute inference. Our experimental results show that contrastive models are less vulnerable to membership inference attacks but more vulnerable to attribute inference attacks compared to supervised models. The former is due to the fact that contrastive models are less prone to overfitting, while the latter is caused by contrastive models' capability of representing data samples expressively. To remedy this situation, we propose the first privacy-preserving contrastive learning mechanism, namely Talos, relying on adversarial training. Empirical results show that Talos can successfully mitigate attribute inference risks for contrastive models while maintaining their membership privacy and model utility.
翻译:在过去十年中,数据是推动机器学习(ML)发展的关键因素。然而,高质量的数据,特别是标签数据,往往很难收集,而且费用昂贵。为了利用大规模无标签数据,引入了以对比性学习为代表的自我监督学习。对比式学习的目的是绘制来自培训样本的不同观点(例如,通过数据增强),其代表空间更接近,而从不同样本中得出的不同观点则更为遥远。通过这种方式,对比性模型学会为数据样本制作信息化演示,然后用于下游 ML任务。最近的研究表明,机器学习模型容易受到各种隐私攻击的伤害。然而,目前大多数工作都集中在通过监督性学习培训的模式上。与此同时,通过对比性学习而获得的数据样本信息性展示可能带来严重的隐私风险。在本文中,我们通过成员推算模型和归因的推论进行第一次对比性分析,我们的实验结果显示,对比性模型对于成员推论攻击的脆弱性较小,但更易受到归属性攻击的影响。与监督性模型相比,分析性测试性模型的精确性模型则显示,前者的精确性分析结果可以显示,而后一种纠正性模型则显示,而导致数据的修复性模型。前一种比较性模型。