Hate speech in social media is a growing phenomenon, and detecting such toxic content has recently gained significant traction in the research community. Existing studies have explored fine-tuning language models (LMs) to perform hate speech detection, and these solutions have yielded significant performance. However, most of these studies are limited to detecting hate speech only in English, neglecting the bulk of hateful content that is generated in other languages, particularly in low-resource languages. Developing a classifier that captures hate speech and nuances in a low-resource language with limited data is extremely challenging. To fill the research gap, we propose HateMAML, a model-agnostic meta-learning-based framework that effectively performs hate speech detection in low-resource languages. HateMAML utilizes a self-supervision strategy to overcome the limitation of data scarcity and produces better LM initialization for fast adaptation to an unseen target language (i.e., cross-lingual transfer) or other hate speech datasets (i.e., domain generalization). Extensive experiments are conducted on five datasets across eight different low-resource languages. The results show that HateMAML outperforms the state-of-the-art baselines by more than 3% in the cross-domain multilingual transfer setting. We also conduct ablation studies to analyze the characteristics of HateMAML.
翻译:社交媒体中的仇恨言论是一个日益增长的现象,发现这种有毒内容最近在研究界已获得显著的吸引力。现有研究探索了微调语言模式,以进行仇恨言论检测,这些解决方案也取得了显著成效。然而,大多数研究仅限于检测仇恨言论,忽视了以其他语言,特别是以低资源语言生成的大量仇恨内容。开发一个在低资源语言中捕捉仇恨言论和细微差别的分类器,而数据有限,这是极具挑战性的。为了填补研究空白,我们建议HateMAML(HateMAML),这是一个基于以低资源语言有效进行仇恨言论检测的示范性、不可知性的元学习框架。HateMAML(HateMAML)利用自我监督战略克服数据稀缺的限制,并产生更好的LMM(LM)初始化,以便快速适应一种隐蔽的目标语言(即跨语言传输)或其他仇恨言论数据集(即广域化)。在八种低资源语言的五套数据集上进行了广泛的实验。结果显示,HateMAML(H)超越了以低资源语言为低资源语言有效进行仇恨言论检测。通过州际3MLML(ML)基准,我们还进行了比多语言分析。</s>