Blind spots or outright deceit can bedevil and deceive machine learning models. Unidentified objects such as digital "stickers," also known as adversarial patches, can fool facial recognition systems, surveillance systems and self-driving cars. Fortunately, most existing adversarial patches can be outwitted, disabled and rejected by a simple classification network called an adversarial patch detector, which distinguishes adversarial patches from original images. An object detector classifies and predicts the types of objects within an image, such as by distinguishing a motorcyclist from the motorcycle, while also localizing each object's placement within the image by "drawing" so-called bounding boxes around each object, once again separating the motorcyclist from the motorcycle. To train detectors even better, however, we need to keep subjecting them to confusing or deceitful adversarial patches as we probe for the models' blind spots. For such probes, we came up with a novel approach, a Low-Detectable Adversarial Patch, which attacks an object detector with small and texture-consistent adversarial patches, making these adversaries less likely to be recognized. Concretely, we use several geometric primitives to model the shapes and positions of the patches. To enhance our attack performance, we also assign different weights to the bounding boxes in terms of loss function. Our experiments on the common detection dataset COCO as well as the driving-video dataset D2-City show that LDAP is an effective attack method, and can resist the adversarial patch detector.
翻译:盲点或直接的欺骗可以诱骗和欺骗机器学习模型。 数字“ 棍棒” 等未知物体( 也称为对抗性板块 ) 可以愚弄面部识别系统、 监视系统和自行驾驶汽车。 幸运的是, 多数现有的对抗性补丁可以被一个叫作对抗性补丁检测器的简单分类网络所取代、 禁用和拒绝, 它将对抗性补丁与原始图像区分开来。 一个物体检测器将图像中的物体类型分类和预测, 比如通过区分摩托车和摩托车区分机动车, 同时将每个物体在图像中的位置定位为“ 拖动” 所谓的捆绑框, 再次将摩托车和摩托车分开。 然而, 幸运的是, 多数现有的对抗性补丁( ) 要更好地训练探测器, 我们需要在模型的盲点中不断将它们置于混乱或欺骗性的对抗性对抗性补丁。 对于这样的探测器, 我们提出了一种新颖的方法, 低可检测性的Adversarial补丁( ) 用来攻击一个小型和有文字相容的对立的对立的对立的对立的对立点, 使得这些对立的对立的对立器更可能更少的对立性定位, 我们的对立性 的对立性数据进行精确的测量, 我们的对质的测测测判的对质的对质的对质的测。