In transfer learning, only the last part of the networks - the so-called head - is often fine-tuned. Representation similarity analysis shows that the most significant change still occurs in the head even if all weights are updatable. However, recent results from few-shot learning have shown that representation change in the early layers, which are mostly convolutional, is beneficial, especially in the case of cross-domain adaption. In our paper, we find out whether that also holds true for transfer learning. In addition, we analyze the change of representation in transfer learning, both during pre-training and fine-tuning, and find out that pre-trained structure is unlearned if not usable.
翻译:在转移学习方面,只有最后一部分网络,即所谓的头部,经常进行微调。代表性相似性分析显示,即使所有重量都可提高,但头部仍发生最重大的变化。然而,最近一些短短的学习结果表明,早期(大多是革命性的)代表性的变化是有益的,特别是在跨部适应方面。我们发现,在我们的论文中,这是否也适用于转移学习。此外,我们分析了在培训前和微调期间转移学习代表性的变化,并发现培训前的结构即使不能使用,也是无法学习的。