Deep learning has been widely used in data-intensive applications. However, training a deep neural network often requires a large data set. When there is not enough data available for training, the performance of deep learning models is even worse than that of shallow networks. It has been proved that few-shot learning can generalize to new tasks with few training samples. Fine-tuning of a deep model is simple and effective few-shot learning method. However, how to fine-tune deep learning models (fine-tune convolution layer or BN layer?) still lack deep investigation. Hence, we study how to fine-tune deep models through experimental comparison in this paper. Furthermore, the weight of the models is analyzed to verify the feasibility of the fine-tuning method.
翻译:深层学习被广泛用于数据密集型应用。然而,深神经网络的培训往往需要大型数据集。当没有足够的培训数据时,深层学习模型的性能甚至比浅网络的性能还要差。实践证明,少见的学习可以用少量培训样本来概括新的任务。深层模型的微调是简单而有效的少见的学习方法。然而,如何微调深层学习模型(松果层或BN层?)仍然缺乏深入的调查。因此,我们研究如何通过本文中的实验性比较来微调深层模型。此外,对模型的权重进行了分析,以核实微调方法的可行性。