Deep neural networks (DNNs) are sensitive and susceptible to tiny perturbation by adversarial attacks which causes erroneous predictions. Various methods, including adversarial defense and uncertainty inference (UI), have been developed in recent years to overcome the adversarial attacks. In this paper, we propose a multi-head uncertainty inference (MH-UI) framework for detecting adversarial attack examples. We adopt a multi-head architecture with multiple prediction heads (i.e., classifiers) to obtain predictions from different depths in the DNNs and introduce shallow information for the UI. Using independent heads at different depths, the normalized predictions are assumed to follow the same Dirichlet distribution, and we estimate distribution parameter of it by moment matching. Cognitive uncertainty brought by the adversarial attacks will be reflected and amplified on the distribution. Experimental results show that the proposed MH-UI framework can outperform all the referred UI methods in the adversarial attack detection task with different settings.
翻译:深海神经网络(DNN)敏感,很容易受到引起错误预测的对抗性攻击的微小扰动。近年来,为克服对抗性攻击,已经开发了各种方法,包括对抗性防御和不确定推论(UI),以克服对抗性攻击。在本文件中,我们提出了一个多头不确定推论(MH-UI)框架,以探测对抗性攻击的例子。我们采用了多头多头结构,有多个预测头(即分类者),以便从DNNM的不同深度获得预测,并为UI引进浅信息。在不同深度使用独立头,假设标准化预测将遵循同样的Drichlet分布,我们估计其分布参数在瞬间匹配。对立性攻击带来的认知不确定性将反映并放大在分布上。实验结果显示,拟议的MH-UI框架可以超越对抗性攻击性攻击探测任务中不同环境的所有提到的UI方法。