Knowledge distillation is a popular paradigm for learning portable neural networks by transferring the knowledge from a large model into a smaller one. Most existing approaches enhance the student model by utilizing the similarity information between the categories of instance level provided by the teacher model. However, these works ignore the similarity correlation between different instances that plays an important role in confidence prediction. To tackle this issue, we propose a novel method in this paper, called similarity transfer for knowledge distillation (STKD), which aims to fully utilize the similarities between categories of multiple samples. Furthermore, we propose to better capture the similarity correlation between different instances by the mixup technique, which creates virtual samples by a weighted linear interpolation. Note that, our distillation loss can fully utilize the incorrect classes similarities by the mixed labels. The proposed approach promotes the performance of student model as the virtual sample created by multiple images produces a similar probability distribution in the teacher and student networks. Experiments and ablation studies on several public classification datasets including CIFAR-10,CIFAR-100,CINIC-10 and Tiny-ImageNet verify that this light-weight method can effectively boost the performance of the compact student model. It shows that STKD substantially has outperformed the vanilla knowledge distillation and has achieved superior accuracy over the state-of-the-art knowledge distillation methods.
翻译:知识蒸馏是学习便携式神经网络的流行范例,通过将一个大模型的知识转换成一个较小的模型来学习便携式神经网络。大多数现有方法都利用教师模型提供的实验级别类别之间的相似信息来增强学生模型。然而,这些方法忽略了在信任预测中发挥重要作用的不同实例之间的相似性关系。为了解决这一问题,我们在本文件中提出一种新的方法,称为知识蒸馏的相似性转移(STKD),目的是充分利用多种样本类别之间的相似性。此外,我们提议更好地通过混合技术来捕捉不同实例之间的相似性相关性,这种混合技术通过加权线性线性内插图生成虚拟样本。指出,我们的蒸馏损失可以充分利用混合标签的不正确的类相似性。拟议方法促进学生模型的性能,因为由多个图像生成的虚拟样本在教师和学生网络中产生类似的概率分布。对几个公共分类数据集进行了实验和对比研究,包括CIFAR-10、CIFAR-100、CINIIC-10和Tiny-ImageNet, 来更好地捕捉摸不同实例,通过加权的线性图象学方法来创造虚拟样本样本样本样本样本样本样本样本样本样本样本样本样本样本样本。注意,我们完全利用混合标签标签标签标签标签标签标签标签标签标签标签标签标签标签标签标签标签标签标签标签标签标签标签标签标签标签标签标签的精质的精质谱的精度,可以提升学生的精准性知识,从而有效地分析。