Hate speech detection is complex; it relies on commonsense reasoning, knowledge of stereotypes, and an understanding of social nuance that differs from one culture to the next. It is also difficult to collect a large-scale hate speech annotated dataset. In this work, we frame this problem as a few-shot learning task, and show significant gains with decomposing the task into its "constituent" parts. In addition, we see that infusing knowledge from reasoning datasets (e.g. Atomic2020) improves the performance even further. Moreover, we observe that the trained models generalize to out-of-distribution datasets, showing the superiority of task decomposition and knowledge infusion compared to previously used methods. Concretely, our method outperforms the baseline by 17.83% absolute gain in the 16-shot case.
翻译:仇恨言论的检测是复杂的;它依赖于常识推理、对陈规定型观念的了解,以及对不同文化的社会细微差别的理解。 收集大规模仇恨言论附加说明的数据集也很困难。 在这项工作中,我们将这一问题描述为一小段学习任务,并显示出将任务分解为“组成”部分所取得的显著成果。 此外,我们看到,从推理数据集(例如原子2020年)中吸收知识会进一步提高业绩。 此外,我们观察到,经过培训的模型将分布数据集普遍化,显示任务分解和知识融合优于先前使用的方法。 具体地说,我们的方法比16个截图案例的基线高出17.83%的绝对收益。