Few-shot learning models learn representations with limited human annotations, and such a learning paradigm demonstrates practicability in various tasks, e.g., image classification, object detection, etc. However, few-shot object detection methods suffer from an intrinsic defect that the limited training data makes the model cannot sufficiently explore semantic information. To tackle this, we introduce knowledge distillation to the few-shot object detection learning paradigm. We further run a motivating experiment, which demonstrates that in the process of knowledge distillation, the empirical error of the teacher model degenerates the prediction performance of the few-shot object detection model as the student. To understand the reasons behind this phenomenon, we revisit the learning paradigm of knowledge distillation on the few-shot object detection task from the causal theoretic standpoint, and accordingly, develop a Structural Causal Model. Following the theoretical guidance, we propose a backdoor adjustment-based knowledge distillation method for the few-shot object detection task, namely Disentangle and Remerge (D&R), to perform conditional causal intervention toward the corresponding Structural Causal Model. Empirically, the experiments on benchmarks demonstrate that D&R can yield significant performance boosts in few-shot object detection. Code is available at https://github.com/ZYN-1101/DandR.git.
翻译:少见的学习模型以有限的人文说明来学习表现形式,而这种学习模式则表明,在各种任务中,例如图像分类、物体探测等,都具有实用性。 然而,少见的物体探测方法存在一个内在缺陷,即有限的培训数据使模型无法充分探索语义信息。要解决这个问题,我们将知识蒸馏引入微小的物体探测学习模式。我们进一步进行激励性实验,表明在知识蒸馏过程中,教师模型的经验错误使作为学生的微小物体探测模型的预测性能退化。为了了解这一现象背后的原因,我们从因果关系角度重新审视微小物体探测任务方面的知识蒸馏学习模式,并据此开发一个结构构造模型。根据理论指导,我们为微小物体探测任务提出了基于后门调整的知识蒸馏方法,即分解和Remerge(D & R&R),以便向相应的结构剖析模型进行有条件的因果干涉性干扰性干预。 精选,基准实验显示,D & R-RDC 能够产生显著的推进性。