Many natural language processing (NLP) tasks are naturally imbalanced, as some target categories occur much more frequently than others in the real world. In such scenarios, current NLP models still tend to perform poorly on less frequent classes. Addressing class imbalance in NLP is an active research topic, yet, finding a good approach for a particular task and imbalance scenario is difficult. With this survey, the first overview on class imbalance in deep-learning based NLP, we provide guidance for NLP researchers and practitioners dealing with imbalanced data. We first discuss various types of controlled and real-world class imbalance. Our survey then covers approaches that have been explicitly proposed for class-imbalanced NLP tasks or, originating in the computer vision community, have been evaluated on them. We organize the methods by whether they are based on sampling, data augmentation, choice of loss function, staged learning, or model design. Finally, we discuss open problems such as dealing with multi-label scenarios, and propose systematic benchmarking and reporting in order to move forward on this problem as a community.
翻译:许多自然语言处理(NLP)任务自然是不平衡的,因为有些目标类别比现实世界中的其他类别更经常出现。在这种情况下,目前的NLP模式仍然倾向于在不常见的班级上表现不佳。解决NLP中的阶级不平衡是一个积极的研究课题。解决NLP中的阶级不平衡问题是一个积极的研究课题,然而,为某一特定任务和不平衡情况寻找一个好的方法是困难的。通过这项调查,我们首次对基于NLP的深层次学习中的阶级不平衡问题进行概述,我们为NLP研究人员和从业人员处理不平衡数据问题提供指导。我们首先讨论各种类型的受控和实际世界级不平衡问题。我们的调查随后涵盖了为班级平衡NLP任务明确提议的办法,或者从计算机视野社区开始,已经对这些办法进行了评估。我们根据这些方法的抽样、数据增强、损失函数的选择、分阶段学习或模型设计来组织方法。最后,我们讨论一些公开的问题,例如处理多标签情景,并提出系统的基准和报告,以便作为一个社区推进这一问题。