The success of deep learning is usually accompanied by the growth in neural network depth. However, the traditional training method only supervises the neural network at its last layer and propagates the supervision layer-by-layer, which leads to hardship in optimizing the intermediate layers. Recently, deep supervision has been proposed to add auxiliary classifiers to the intermediate layers of deep neural networks. By optimizing these auxiliary classifiers with the supervised task loss, the supervision can be applied to the shallow layers directly. However, deep supervision conflicts with the well-known observation that the shallow layers learn low-level features instead of task-biased high-level semantic features. To address this issue, this paper proposes a novel training framework named Contrastive Deep Supervision, which supervises the intermediate layers with augmentation-based contrastive learning. Experimental results on nine popular datasets with eleven models demonstrate its effects on general image classification, fine-grained image classification and object detection in supervised learning, semi-supervised learning and knowledge distillation. Codes have been released in Github.
翻译:深层学习的成功通常伴随着神经网络深度的增长,然而,传统培训方法只监督最后一层的神经网络,并逐层传播监督层,这导致优化中间层的困难。最近,提出了在深层神经网络中间层中增加辅助分类器的建议。通过优化这些辅助分类器,在监督任务损失的情况下,监督可直接应用于浅层。但是,深层监督与众所周知的观测相冲突,即浅层学习低层次特征,而不是高层次任务偏重的语义特征。为解决这一问题,本文提出了一个名为“反向深层监督”的新培训框架,以扩大的对比性学习来监督中间层。九种流行数据集的实验结果和十一种模型显示了其对一般图像分类、精细微图像分类和在监督学习、半监督学习和知识提炼过程中的物体探测效果。在吉舒布发布了代码。