Domain adaptive object detection (DAOD) aims to alleviate transfer performance degradation caused by the cross-domain discrepancy. However, most existing DAOD methods are dominated by computationally intensive two-stage detectors, which are not the first choice for industrial applications. In this paper, we propose a novel semi-supervised domain adaptive YOLO (SSDA-YOLO) based method to improve cross-domain detection performance by integrating the compact one-stage detector YOLOv5 with domain adaptation. Specifically, we adapt the knowledge distillation framework with the Mean Teacher model to assist the student model in obtaining instance-level features of the unlabeled target domain. We also utilize the scene style transfer to cross-generate pseudo images in different domains for remedying image-level differences. In addition, an intuitive consistency loss is proposed to further align cross-domain predictions. We evaluate our proposed SSDA-YOLO on public benchmarks including PascalVOC, Clipart1k, Cityscapes, and Foggy Cityscapes. Moreover, to verify its generalization, we conduct experiments on yawning detection datasets collected from various classrooms. The results show considerable improvements of our method in these DAOD tasks. Our code is available on \url{https://github.com/hnuzhy/SSDA-YOLO}.
翻译:内地适应性天体探测(DAOD)旨在缓解跨域差异造成的转移性性能退化,然而,大多数现有的DAOD方法主要是以计算密集的两阶段探测器为主,这不是工业应用的首选。在本文件中,我们提议了一种新的半监督域适应性YOLO(SSDA-YOLO) (SDA-YOLO) (SDA-YOLO) 基础方法,通过将小型一阶段探测器YOLOv5 (YOLOv5) 与域适应性能相结合,改进跨域性能探测性能。具体地说,我们将知识蒸馏框架与普通教师模型相适应,以协助学生模型获得未标目标域的试级特征。我们还利用场景风格转换到不同领域的跨基因假象图像,以补救图像级差异。此外,还提议了一种直观的一致性损失,以进一步调整跨域预测。我们提议的SDADA-YOL-YOLOO(C)、Clipalartart1k、城市景象、Foggy Cases,以及Foggy Cas, 校区景模型实验,我们收集的自动检测/DADADADADODO/DODADR) 的可相当的系统。