Building advanced machine learning (ML) models requires expert knowledge and many trials to discover the best architecture and hyperparameter settings. Previous work demonstrates that model information can be leveraged to assist other attacks, such as membership inference, generating adversarial examples. Therefore, such information, e.g., hyperparameters, should be kept confidential. It is well known that an adversary can leverage a target ML model's output to steal the model's information. In this paper, we discover a new side channel for model information stealing attacks, i.e., models' scientific plots which are extensively used to demonstrate model performance and are easily accessible. Our attack is simple and straightforward. We leverage the shadow model training techniques to generate training data for the attack model which is essentially an image classifier. Extensive evaluation on three benchmark datasets shows that our proposed attack can effectively infer the architecture/hyperparameters of image classifiers based on convolutional neural network (CNN) given the scientific plot generated from it. We also reveal that the attack's success is mainly caused by the shape of the scientific plots, and further demonstrate that the attacks are robust in various scenarios. Given the simplicity and effectiveness of the attack method, our study indicates scientific plots indeed constitute a valid side channel for model information stealing attacks. To mitigate the attacks, we propose several defense mechanisms that can reduce the original attacks' accuracy while maintaining the plot utility. However, such defenses can still be bypassed by adaptive attacks.
翻译:建立先进的机器学习(ML)模型需要专家知识和许多试验才能发现最佳建筑和超参数设置。 先前的工作表明,模型信息可以用来协助其他攻击,例如会籍推断,产生对抗性实例。 因此,这类信息,例如超参数,应该保密。 众所周知,一个对手可以利用目标ML模型输出来窃取模型信息。 在本文中,我们发现了一个用于模式信息盗窃袭击的新侧渠道,即广泛用来展示模型性能和容易获得的科学图案。 我们的攻击是简单和直截了当的。 我们利用影子模型培训技术来为攻击模型(基本上是一个图像分类器)生成培训数据。 对三个基准数据集的广泛评估表明,我们拟议的攻击可以有效地推导出基于革命神经网络(CNN)的图像分类的架构/功能。 我们还可以发现,攻击的成功主要来自科学图的形状,我们的攻击是简单化的模型,我们进一步证明攻击是准确性的。 我们的系统攻击是精确的。 精确度,我们为各种攻击提供了一种精确性攻击方法。