In recent years, binary analysis gained traction as a fundamental approach to inspect software and guarantee its security. Due to the exponential increase of devices running software, much research is now moving towards new autonomous solutions based on deep learning models, as they have been showing state-of-the-art performances in solving binary analysis problems. One of the hot topics in this context is binary similarity, which consists in determining if two functions in assembly code are compiled from the same source code. However, it is unclear how deep learning models for binary similarity behave in an adversarial context. In this paper, we study the resilience of binary similarity models against adversarial examples, showing that they are susceptible to both targeted and untargeted attacks (w.r.t. similarity goals) performed by black-box and white-box attackers. In more detail, we extensively test three current state-of-the-art solutions for binary similarity against two black-box greedy attacks, including a new technique that we call Spatial Greedy, and one white-box attack in which we repurpose a gradient-guided strategy used in attacks to image classifiers.
翻译:近年来,二进制分析作为检测软件并确保其安全性的一种基本方法得到了广泛关注。随着运行软件设备的呈指数增长,许多研究现在正在转向基于深度学习模型的新型自主解决方案,因为它们在解决二进制分析问题方面表现出最先进的性能。其中一个热门主题是二进制相似性,它包括确定两个汇编代码中的函数是否来自同一源代码。然而,深度学习模型在对抗性背景下的二进制相似性行为尚未明确。在本文中,我们研究了二进制相似性模型对抗性示例的抵抗力,展示了它们容易受到白盒和黑盒攻击者执行的面向相似性目标的有针对性和无针对性攻击。具体地,我们广泛测试了三种当前的二进制相似性解决方案,用两种黑盒贪婪攻击(包括我们称为空间贪婪的新技术)和一种白盒攻击,在这种攻击中,我们重新利用了一种攻击图像分类器的梯度引导策略。