In recent years, deep learning has shown performance breakthroughs in many applications, such as image detection, image segmentation, pose estimation, and speech recognition. However, this comes with a major concern: deep networks have been found to be vulnerable to adversarial examples. Adversarial examples are slightly modified inputs that are intentionally designed to cause a misclassification by the model. In the domains of images and speech, the modifications are so small that they are not seen or heard by humans, but nevertheless greatly affect the classification of the model. Deep learning models have been successfully applied to malware detection. In this domain, generating adversarial examples is not straightforward, as small modifications to the bytes of the file could lead to significant changes in its functionality and validity. We introduce a novel loss function for generating adversarial examples specifically tailored for discrete input sets, such as executable bytes. We modify malicious binaries so that they would be detected as benign, while preserving their original functionality, by injecting a small sequence of bytes (payload) in the binary file. We applied this approach to an end-to-end convolutional deep learning malware detection model and show a high rate of detection evasion. Moreover, we show that our generated payload is robust enough to be transferable within different locations of the same file and across different files, and that its entropy is low and similar to that of benign data sections.
翻译:近些年来,深层次的学习显示许多应用软件的绩效突破,如图像检测、图像分割、显示估计和语音识别等。然而,这引起一个重大关切:深层次的网络被发现容易发生对抗性实例。反向实例是有意设计的一些略有修改的投入,目的是造成模型的分类错误。在图像和语言方面,这些修改太小,以至于人类看不到或听不到模型的分类,但大大影响了模型的分类。深层次的学习模型被成功地应用于恶意软件的检测。在这一领域,生成对抗性实例并非直截了当,因为对文件的字节进行小小的修改可能会导致其功能和有效性的重大变化。我们引入了一种新的损失功能,用于生成专门针对独立输入数据集的对抗性实例,例如可执行性。我们在图像和语言方面修改恶意的双曲,以保持其原始的功能,在二进制文档中注入一小串的字节(顶部) 。我们将这一方法应用到一个端端端端端端深层次的深层学习恶意检测模型和有效性的细小节可以导致其功能和有效性的显著改变。此外,我们所生成的低版本的版本的快速检测位置也展示了它所生成的高度的版本,从而显示的高级的低层数据。