In recent years, the widespread use of social media has led to an increase in the generation of toxic and offensive content on online platforms. In response, social media platforms have worked on developing automatic detection methods and employing human moderators to cope with this deluge of offensive content. While various state-of-the-art statistical models have been applied to detect toxic posts, there are only a few studies that focus on detecting the words or expressions that make a post offensive. This motivates the organization of the SemEval-2021 Task 5: Toxic Spans Detection competition, which has provided participants with a dataset containing toxic spans annotation in English posts. In this paper, we present the WLV-RIT entry for the SemEval-2021 Task 5. Our best performing neural transformer model achieves an $0.68$ F1-Score. Furthermore, we develop an open-source framework for multilingual detection of offensive spans, i.e., MUDES, based on neural transformers that detect toxic spans in texts.
翻译:近年来,广泛使用社交媒体导致在线平台生成有毒和攻击性内容的数量增多,社交媒体平台为此开发了自动检测方法,并聘请了人类主持人应对这种大量攻击性内容,虽然应用了各种最先进的统计模型来检测有毒内容,但只有几项研究侧重于检测作出攻击性文章的文字或表达方式,这促使组织了SemEval-2021任务5:有毒螺旋探测竞赛,为参与者提供了含有有毒内容的英文文章注解数据集。在本文件中,我们为SemEval-2021任务5展示了WLV-RIT条目。我们最出色的神经变压器模型实现了0.68美元F1-Score。此外,我们开发了一个用于多语种探测攻击性空间的开放源框架,即MUDES,其基础是能够检测文本中有毒内容的神经变压器。