Object detection is the foundation of various critical computer-vision tasks such as segmentation, object tracking, and event detection. To train an object detector with satisfactory accuracy, a large amount of data is required. However, due to the intensive workforce involved with annotating large datasets, such a data curation task is often outsourced to a third party or relied on volunteers. This work reveals severe vulnerabilities of such data curation pipeline. We propose MACAB that crafts clean-annotated images to stealthily implant the backdoor into the object detectors trained on them even when the data curator can manually audit the images. We observe that the backdoor effect of both misclassification and the cloaking are robustly achieved in the wild when the backdoor is activated with inconspicuously natural physical triggers. Backdooring non-classification object detection with clean-annotation is challenging compared to backdooring existing image classification tasks with clean-label, owing to the complexity of having multiple objects within each frame, including victim and non-victim objects. The efficacy of the MACAB is ensured by constructively i abusing the image-scaling function used by the deep learning framework, ii incorporating the proposed adversarial clean image replica technique, and iii combining poison data selection criteria given constrained attacking budget. Extensive experiments demonstrate that MACAB exhibits more than 90% attack success rate under various real-world scenes. This includes both cloaking and misclassification backdoor effect even restricted with a small attack budget. The poisoned samples cannot be effectively identified by state-of-the-art detection techniques.The comprehensive video demo is at https://youtu.be/MA7L_LpXkp4, which is based on a poison rate of 0.14% for YOLOv4 cloaking backdoor and Faster R-CNN misclassification backdoor.
翻译:目标检测是各种关键计算机视野任务的基础, 如分割、 对象跟踪和事件检测。 要对对象探测器进行准确无误的培训, 需要大量数据。 但是, 由于大量员工参与批注大型数据集, 此类数据曲线化任务往往外包给第三方或依靠志愿者。 这项工作揭示了这些数据曲线化管道的严重脆弱性。 我们建议, 翻译干净的附加说明的图像, 以便悄悄地将后门植入经过培训的物体探测器, 即使数据管理员可以手动对图像进行审计。 我们观察到, 错误分类化和隐蔽两者的后门效应都是在野外强有力地实现的。 当后门以不明显的自然触发器启动时, 此类数据曲线化任务往往外包给第三方, 或依赖志愿者。 我们建议, 制作干净的后门式图像, 包括受害者和非受害者物体。 通过建设性的方式, 将图像缩放的后门化和隐蔽式样本化的后端功能 。 在深度测试框架下, 将不易变形的图像缩缩缩缩 。