Deep learning usually relies on training large-scale data samples to achieve better performance. However, over-fitting based on training data always remains a problem. Scholars have proposed various strategies, such as feature dropping and feature mixing, to improve the generalization continuously. For the same purpose, we subversively propose a novel training method, Feature Weaken, which can be regarded as a data augmentation method. Feature Weaken constructs the vicinal data distribution with the same cosine similarity for model training by weakening features of the original samples. In especially, Feature Weaken changes the spatial distribution of samples, adjusts sample boundaries, and reduces the gradient optimization value of back-propagation. This work can not only improve the classification performance and generalization of the model, but also stabilize the model training and accelerate the model convergence. We conduct extensive experiments on classical deep convolution neural models with five common image classification datasets and the Bert model with four common text classification datasets. Compared with the classical models or the generalization improvement methods, such as Dropout, Mixup, Cutout, and CutMix, Feature Weaken shows good compatibility and performance. We also use adversarial samples to perform the robustness experiments, and the results show that Feature Weaken is effective in improving the robustness of the model.
翻译:深层学习通常依赖于培训大型数据样本,以取得更好的性能。然而,基于培训数据的过度搭配始终是一个问题。学者们提出了各种战略,如特征下降和特征混合等,以不断改进一般化。为了同样的目的,我们颠覆性地提议一种新的培训方法,即地貌Weaken,可被视为一种数据增强方法。功能Weaken通过削弱原始样本的特征,为模型培训构建了相似的矩形数据分布。特别是,地貌Weaken改变了样本的空间分布,调整了样本的边界,并降低了后方转换的梯度优化值。这项工作不仅可以改进模型的分类性能和一般化,而且还可以稳定模型培训并加速模型的趋同。我们用五种通用图像分类数据集对古典的卷动神经模型进行广泛的实验,用四种通用文本分类数据集对伯特模型进行广泛的实验。与古典模型或一般化改进方法相比,例如脱落、Mixup、Cutout、CutMix、Feart-Mix、Wetary weken Weken wekent weken 实验也展示了可靠的兼容性和性实验,我们展示了强性。我们也展示了对立性地展示了良好的性结果。