DETR is a novel end-to-end transformer architecture object detector, which significantly outperforms classic detectors when scaling up the model size. In this paper, we focus on the compression of DETR with knowledge distillation. While knowledge distillation has been well-studied in classic detectors, there is a lack of researches on how to make it work effectively on DETR. We first provide experimental and theoretical analysis to point out that the main challenge in DETR distillation is the lack of consistent distillation points. Distillation points refer to the corresponding inputs of the predictions for student to mimic, and reliable distillation requires sufficient distillation points which are consistent between teacher and student. Based on this observation, we propose a general knowledge distillation paradigm for DETR(KD-DETR) with consistent distillation points sampling. Specifically, we decouple detection and distillation tasks by introducing a set of specialized object queries to construct distillation points. In this paradigm, we further propose a general-to-specific distillation points sampling strategy to explore the extensibility of KD-DETR. Extensive experiments on different DETR architectures with various scales of backbones and transformer layers validate the effectiveness and generalization of KD-DETR. KD-DETR boosts the performance of DAB-DETR with ResNet-18 and ResNet-50 backbone to 41.4$\%$, 45.7$\%$ mAP, respectively, which are 5.2$\%$, 3.5$\%$ higher than the baseline, and ResNet-50 even surpasses the teacher model by $2.2\%$.
翻译:DETR是一个新的端到端变压器结构物体探测器,在扩大模型规模时,它明显优于经典探测器。在本文中,我们侧重于通过知识蒸馏压缩DETR。虽然知识蒸馏在经典探测器中研究过,但缺乏关于如何使DETR(KD-DETR)有效运行的研究。我们首先提供实验和理论分析,指出DETR蒸馏的主要挑战是缺乏一致的蒸馏点。蒸馏点是指学生模拟预测的相应投入,可靠的蒸馏需要足够的蒸馏点,这与师生之间是一致的。根据这一观察,我们提出了DTR(KD-DETR)一般蒸馏点取样的普通知识蒸馏模式。我们首先提供一套专门的对象查询,以构建蒸馏点。在这个模型中,我们进一步建议一个通用的蒸馏点对学生进行模拟预测的对应投入,而可靠的蒸馏点需要有足够的蒸馏点,而教师和学生之间的蒸馏点需要有足够的蒸馏点。根据这一观察,我们建议为DETR(K-DE-DE-DE-DE-DE-DE-DE-DE-DE-Q-L-S-S-Sl-S-Syal-Slal-Slal-S-S-S-Syal-Syal-Slationxxxxxxxxxxxxxxxxxxxxxxxxxxx 和S-DE-DE-DE-DE-DE-DE-S-S-S-S-S-D-S-S-S-S-S-S-S-S-S-S-D-S-S-D-S-S-S-S-D-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-D-D-Sl-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-