Adversarial attacks on state-of-the-art machine learning models pose a significant threat to the safety and security of mission-critical autonomous systems. This paper considers the additional vulnerability of machine learning models when attackers can measure the power consumption of their underlying hardware platform. In particular, we explore the utility of power consumption information for adversarial attacks on non-volatile memory crossbar-based single-layer neural networks. Our results from experiments with MNIST and CIFAR-10 datasets show that power consumption can reveal important information about the neural network's weight matrix, such as the 1-norm of its columns. That information can be used to infer the sensitivity of the network's loss with respect to different inputs. We also find that surrogate-based black box attacks that utilize crossbar power information can lead to improved attack efficiency.
翻译:对最先进的机器学习模型的反向攻击对任务关键自主系统的安全和安保构成重大威胁。本文件考虑了当攻击者能够测量其基本硬件平台的电耗时,机器学习模型的额外脆弱性。特别是,我们探讨了动力消耗信息对非挥发性内存跨截线的单层神经网络进行对抗性攻击的效用。我们对MNIST和CIFAR-10数据集的实验结果表明,电力消耗可以揭示神经网络重量矩阵的重要信息,例如其一栏的中枢。这种信息可用来推断网络在不同投入方面损失的敏感性。我们还发现,利用跨管电源信息的代用黑盒攻击可以提高攻击效率。