Transfer learning has been widely studied and gained increasing popularity to improve the accuracy of machine learning models by transferring some knowledge acquired in different training. However, no prior work has pointed out that transfer learning can strengthen privacy attacks on machine learning models. In this paper, we propose TransMIA (Transfer learning-based Membership Inference Attacks), which use transfer learning to perform membership inference attacks on the source model when the adversary is able to access the parameters of the transferred model. In particular, we propose a transfer shadow training technique, where an adversary employs the parameters of the transferred model to construct shadow models, to significantly improve the performance of membership inference when a limited amount of shadow training data is available to the adversary. We evaluate our attacks using two real datasets, and show that our attacks outperform the state-of-the-art that does not use our transfer shadow training technique. We also compare four combinations of the learning-based/entropy-based approach and the fine-tuning/freezing approach, all of which employ our transfer shadow training technique. Then we examine the performance of these four approaches based on the distributions of confidence values, and discuss possible countermeasures against our attacks.
翻译:已广泛研究转让学习,并越来越受欢迎,通过转让在不同培训中获得的一些知识来提高机器学习模式的准确性,然而,以前没有一项工作指出,转让学习能够加强对机器学习模式的隐私攻击;在本论文中,我们提议使用转让学习(基于学习的会籍推断攻击)来利用转让学习(基于渗透)模式的参数,对源模式进行成员推断攻击;特别是,我们提议采用转让影子培训技术,让对手利用转让模式的参数来建立影子模型,以大大改进成员推断工作,如果敌人掌握有限的影子培训数据,则可以大大改进成员推断工作;我们利用两个真实的数据集来评估我们的攻击,并表明我们的攻击超出了不使用转让影子培训技术的状态;我们还比较了基于学习(基于渗透)方法与微调/冻结方法的四种组合,所有这些组合都使用了我们的转移影子培训技术;然后,我们根据信任值的分配情况,审视这四种方法的绩效,并讨论针对我们的攻击可能采取的反措施。