Hate speech detection or offensive language detection are well-established but controversial NLP tasks. With 'hate speech' not being a legal term, these tasks often elaborate on the question of which statements are perceived as being 'appropriate' or 'offensive'. Looking beyond this cursory understanding, this article proposes a way to operationalize the legal concept of incitement to hatred as an NLP task. We pursue this task based on the criminal offense of incitement to hatred in {\S} 130 of the German Criminal Code along with the underlying EU Framework. Under the German Network Enforcement Act, social media providers are subject to a direct obligation to delete postings violating this offense. We take this as a use case to study the transition from the ill-defined concepts of hate speech or offensive language which are usually used in NLP to an operationalization of an actual legally binding obligation. We first translate the legal assessment into a series of binary decisions and then collect, annotate, and analyze a dataset according to our annotation scheme. Finally, we translate each of the legal decisions into an NLP task based on the annotated data. The two subtasks of target group detection and targeting act detection can be considered crucial from a legal viewpoint. We show that both can be annotated by non-legally trained persons with sufficient reliability.
翻译:“仇恨言论”不是法律术语,这些任务往往阐述哪些言论被视为“适当”或“攻击性”的问题。除了这一粗略理解之外,这一条还提出了将煽动仇恨这一法律概念作为国家语言方案任务的一种方法。我们根据《德国刑法》第130条煽动仇恨的刑事犯罪以及欧盟基本框架,执行这项任务。根据《德国网络执行法》,社会媒体提供者直接有义务删除违反这一罪行的文章。我们以此为例研究仇恨言论或攻击性语言这一定义错误的概念向实际具有法律约束力的义务的操作过渡。我们首先将法律评估转化为一系列二进制决定,然后根据我们的注释计划收集、注解和分析一个数据集。最后,我们根据德国网络执行法,将每一项法律决定转化为基于附加说明的数据的NLP任务。我们用这个案例来研究从国家语言通常用于国家语言网站的仇恨言论或攻击性语言这一定义不当的概念向实际具有法律约束力的义务的操作过渡。我们先将法律评估转化为一系列的二进制决定,然后根据我们的说明计划收集、注和分析一个数据集。我们可将每项法律决定转化为注释的数据。我们所训练的两子目标都能够显示一个可靠的团体探测和行为,从法律检查的足够可靠的观点。