The loss function is a key component in deep learning models. A commonly used loss function for classification is the cross entropy loss, which is a simple yet effective application of information theory for classification problems. Based on this loss, many other loss functions have been proposed,~\emph{e.g.}, by adding intra-class and inter-class constraints to enhance the discriminative ability of the learned features. However, these loss functions fail to consider the connections between the feature distribution and the model structure. Aiming at addressing this problem, we propose a channel correlation loss (CC-Loss) that is able to constrain the specific relations between classes and channels as well as maintain the intra-class and the inter-class separability. CC-Loss uses a channel attention module to generate channel attention of features for each sample in the training stage. Next, an Euclidean distance matrix is calculated to make the channel attention vectors associated with the same class become identical and to increase the difference between different classes. Finally, we obtain a feature embedding with good intra-class compactness and inter-class separability.Experimental results show that two different backbone models trained with the proposed CC-Loss outperform the state-of-the-art loss functions on three image classification datasets.
翻译:损失函数是深层学习模型中的一个关键组成部分。 通常用于分类的一个损失函数是交叉酶损失,这是对分类问题信息理论的简单而有效的应用。 基于这一损失,提出了许多其他损失函数, ⁇ emph{ e. e. }, 增加了阶级内部和阶级之间的限制, 以提高学习特征的歧视性能力。 但是, 这些损失函数没有考虑到特征分布与模型结构之间的联系。 为了解决这一问题, 我们提议了一个频道相关损失( CC- Los), 能够限制类别和渠道之间的特定关系, 并维持阶级内部和阶级之间的分离性。 CC- Loss 使用一个频道关注模块, 以引起对培训阶段中每个样本特征的注意。 下一步, 计算出一个 Euclidean 远程矩阵, 使与同一类别相关的频道关注矢量变得相同, 并增加不同类别之间的差异。 最后, 我们获得了一个能够限制分类和渠道之间特定关系的频道相关损失( CC- LOS- sparable), 以及维持阶级内部和阶级之间的分离性。 分析结果显示两个不同的骨架模型, 显示两个经过培训的LE- sma 这样的模型, laction 。