With the epidemic continuing, hatred against Asians is intensifying in countries outside Asia, especially among the Chinese. Thus, there is an urgent need to detect and prevent hate speech toward Asians effectively. In this work, we first create COVID-HATE-2022, an annotated dataset that is an extension of the anti-Asian hate speech dataset on Twitter, including 2,035 annotated tweets fetched in early February 2022, which are labeled based on specific criteria, and we present the comprehensive collection of scenarios of hate and non-hate tweets in the dataset. Second, we fine-tune the BERT models based on the relevant datasets, and demonstrate strategies including 1) cleaning the hashtags, usernames being @, URLs, and emojis before the fine-tuning process, and 2) training with the data while validating with the "clean" data (and the opposite) are not effective for improving performance. Third, we investigate the performance of advanced fine-tuning strategies with 1) model-centric approaches, such as discriminative fine-tuning, gradual unfreezing, and warmup steps, and 2) data-centric approaches, which incorporate data trimming and data augmenting, and show that both strategies generally improve the performance, while data-centric ones outperform the others, which demonstrate the feasibility and effectiveness of the data-centric approaches.
翻译:随着这一流行病的继续,亚洲以外的国家,特别是中国,对亚洲人的仇恨正在加剧。因此,迫切需要有效地发现和防止针对亚洲人的仇恨言论。在这项工作中,我们首先创建了COVID-HATE-2022,这是一个附加说明的数据集,这是在推特上扩大反亚洲仇恨言论数据集的延伸,包括2022年2月初在特定标准基础上贴上标签的2,035条附加说明的推文。我们介绍了数据集中仇恨和非仇恨推文的全面收集情况。第二,我们根据相关数据集对BERT模型进行了微调,并展示了各种战略,包括:(1) 清理标签、用户名称是@、 URL和微调过程前的模版,以及(2) 数据培训在与“清洁”数据(和相反)相验证的同时,对改进绩效的效果不起作用。第三,我们调查先进微调战略的绩效,1) 以模型为中心的方法,例如有区别性的微调、逐步解和暖暖步骤,以及展示了各种战略,包括:(1) 清理标签、用户名称是@、 URL和模本方法,同时展示了数据中心方法的绩效,同时展示了数据核心方法,并全面展示了数据核心方法。