Recently published attacks against deep neural networks (DNNs) have stressed the importance of methodologies and tools to assess the security risks of using this technology in critical systems. Efficient techniques for detecting adversarial machine learning helps establishing trust and boost the adoption of deep learning in sensitive and security systems. In this paper, we propose a new technique for defending deep neural network classifiers, and convolutional ones in particular. Our defense is cheap in the sense that it requires less computation power despite a small cost to pay in terms of detection accuracy. The work refers to a recently published technique called ML-LOO. We replace the costly pixel by pixel leave-one-out approach of ML-LOO by adopting coarse-grained leave-one-out. We evaluate and compare the efficiency of different segmentation algorithms for this task. Our results show that a large gain in efficiency is possible, even though penalized by a marginal decrease in detection accuracy.
翻译:最近公布的对深神经网络的袭击强调了评估在关键系统中使用这一技术的安全风险的方法和工具的重要性。检测对抗性机器学习的有效技术有助于在敏感和安全系统中建立信任和推动采用深层学习。在本文中,我们提出了保护深神经网络分类器的新技术,特别是革命性技术。我们的防御成本低,因为它在检测准确性方面需要较少的计算能力,尽管在检测准确性方面需要支付少量费用。工作指的是最近出版的名为ML-LOO的技术。我们采用粗糙的分解算法来评估并比较这项任务的不同分解算法的效率,从而取代了高成本的ML-LOO的像素。我们的结果表明,尽管检测准确性受到边际下降的制约,但效率的大幅提高是可能的。