Pictures of everyday life are inherently multi-label in nature. Hence, multi-label classification is commonly used to analyze their content. In typical multi-label datasets, each picture contains only a few positive labels, and many negative ones. This positive-negative imbalance can result in under-emphasizing gradients from positive labels during training, leading to poor accuracy. In this paper, we introduce a novel asymmetric loss ("ASL"), that operates differently on positive and negative samples. The loss dynamically down-weights the importance of easy negative samples, causing the optimization process to focus more on the positive samples, and also enables to discard mislabeled negative samples. We demonstrate how ASL leads to a more "balanced" network, with increased average probabilities for positive samples, and show how this balanced network is translated to better mAP scores, compared to commonly used losses. Furthermore, we offer a method that can dynamically adjust the level of asymmetry throughout the training. With ASL, we reach new state-of-the-art results on three common multi-label datasets, including achieving 86.6% on MS-COCO. We also demonstrate ASL applicability for other tasks such as fine-grain single-label classification and object detection. ASL is effective, easy to implement, and does not increase the training time or complexity. Implementation is available at: https://github.com/Alibaba-MIIL/ASL.
翻译:日常生活的图片本质上是多标签性质的。 因此, 多标签分类通常用于分析其内容。 在典型的多标签数据集中, 每张照片只包含几个正面标签和许多负面标签。 这种正偏向性不平衡可能导致在培训期间从正面标签中低估梯度,导致准确性差。 在本文中, 我们引入了一种新的不对称损失( “ ASL ”), 它在正反抽样上运行方式不同。 简单负面样本的重要性是动态的, 导致优化过程更多地关注正样, 并且能够丢弃错误标签的负面样本。 我们展示了ASL如何导致一个更“ 平衡” 的网络, 正面样本的平均概率增加, 并展示了如何将这种平衡网络转化为更好的 mAP 分数, 与通常使用的损失相比。 此外, 我们提供了一种能够动态调整整个培训中不对称程度的方法。 随着 ASLA, 我们到达了三种通用多标签数据集的新状态, 包括达到86.6%的正标, 并且能够丢弃错误标签的样本。 我们还展示了AS- AS- 的可操作性: 在 MS- AS- ASLLALA 和 malalalalal 上, 这样的应用性 A.