DNNs are known to be vulnerable to so-called adversarial attacks, in which inputs are carefully manipulated to induce misclassification. Existing defenses are mostly software-based and come with high overheads or other limitations. This paper presents HASI, a hardware-accelerated defense that uses a process we call stochastic inference to detect adversarial inputs. HASI carefully injects noise into the model at inference time and used the model's response to differentiate adversarial inputs from benign ones. We show an adversarial detection rate of average 87% which exceeds the detection rate of the state-of-the-art approaches, with a much lower overhead. We demonstrate a software/hardware-accelerated co-design, which reduces the performance impact of stochastic inference to 1.58X-2X relative to the unprotected baseline, compared to 14X-20X overhead for a software-only GPU implementation.
翻译:已知DNNs很容易受到所谓的对抗性攻击,在这种攻击中,投入被仔细操纵,以诱导错误分类。现有的防御手段大多以软件为基础,并且具有高管理费或其他限制。本文展示了HASI, 一种硬件加速防御手段,它使用我们称之为随机推论的过程来探测对抗性输入物。我小心地将噪音注入模型中推论时间的模型,并利用模型的反应来区分对抗性输入物和良性输入物。我们显示了平均87 % 的对抗性探测率,它比最新方法的探测率高得多,管理费也低得多。我们展示了软件/硬件加速联合设计,它将随机推论与无防护基线相比的性能影响降低到1.58X-2X,而只使用软件的GPU的14X-20X管理费则降低到14X-20X。