How can citizens moderate hate, toxicity, and extremism in online discourse? We analyze a large corpus of more than 130,000 discussions on German Twitter over the turbulent four years marked by the migrant crisis and political upheavals. With a help of human annotators, language models, machine learning classifiers, and longitudinal statistical analyses, we discern the dynamics of different dimensions of discourse. We find that expressing simple opinions, not necessarily supported by facts but also without insults, relates to the least hate, toxicity, and extremity of speech and speakers in subsequent discussions. Sarcasm also helps in achieving those outcomes, in particular in the presence of organized extreme groups. More constructive comments such as providing facts or exposing contradictions can backfire and attract more extremity. Mentioning either outgroups or ingroups is typically related to a deterioration of discourse in the long run. A pronounced emotional tone, either negative such as anger or fear, or positive such as enthusiasm and pride, also leads to worse outcomes. Going beyond one-shot analyses on smaller samples of discourse, our findings have implications for the successful management of online commons through collective civic moderation.
翻译:公民如何在在线话语中管理仇恨、毒性和极端行为?我们分析了德国Twitter上超过130,000条讨论的大量语料库,在移民危机和政治动荡的四年中进行了分析。借助人工注释器、语言模型、机器学习分类器和纵向统计分析,我们辨别出了不同维度话语的动态。我们发现,表达简单的观点(不一定要有事实支持,但也不侮辱他人),与随后的讨论中言论和演讲者的仇恨、毒性和极端行为最不相关。讽刺在实现这些结果方面也很有帮助,特别是在有组织的极端团体存在的情况下。更具建设性的评论,如提供事实或揭示矛盾,可能会适得其反,吸引更多的极端行为。提及外团体或内团体通常与长期恶化的话语有关。明显的情感语气,无论是消极的愤怒或恐惧,还是积极的热情和自豪感,也会导致更糟糕的结果。超越在较小的话语样本上的一次性分析,我们的发现对通过集体公民管理成功地管理在线共享资源具有重要意义。