Hate speech classification has been a long-standing problem in natural language processing. However, even though there are numerous hate speech detection methods, they usually overlook a lot of hateful statements due to them being implicit in nature. Developing datasets to aid in the task of implicit hate speech classification comes with its own challenges; difficulties are nuances in language, varying definitions of what constitutes hate speech, and the labor-intensive process of annotating such data. This had led to a scarcity of data available to train and test such systems, which gives rise to high variance problems when parameter-heavy transformer-based models are used to address the problem. In this paper, we explore various optimization and regularization techniques and develop a novel RoBERTa-based model that achieves state-of-the-art performance.
翻译:仇恨言论分类在自然语言处理中是一个长期存在的问题,然而,尽管存在许多仇恨言论检测方法,但通常忽视许多仇恨性言论,因为它们具有隐含的本质。 开发数据集以协助隐含仇恨言论分类的任务,这本身也带来了挑战;困难在于语言的细微差别、仇恨言论的不同定义以及这类数据的批注劳动密集型过程。 这导致缺乏可用于培训和测试这类系统的数据,当使用参数重的变压器模型来解决这一问题时,这造成了很大的差异问题。 在本文中,我们探索了各种优化和正规化技术,并开发了新颖的RoBERTA模型,该模型实现了最先进的性能。