Much research has been done for debunking and analysing fake news. Many researchers study fake news detection in the last year, but many are limited to social media data. Currently, multiples fact-checkers are publishing their results in various formats. Also, multiple fact-checkers use different labels for the fake news, making it difficult to make a generalisable classifier. With the merge classes, the performance of the machine model can be enhanced. This domain categorisation will help group the article, which will help save the manual effort in assigning the claim verification. In this paper, we have presented BERT based classification model to predict the domain and classification. We have also used additional data from fact-checked articles. We have achieved a macro F1 score of 83.76 % for Task 3Aand 85.55 % for Task 3B using the additional training data.
翻译:许多研究人员去年研究了假冒新闻探测,但许多只局限于社交媒体数据。目前,多个事实审查员以不同格式公布其结果。此外,多个事实审查员对假新闻使用不同的标签,使得难以形成一个通用分类器。随着合并的分类,机器模型的性能可以得到加强。这个域分类将有助于对文章进行分组,这有助于在分配索赔核实方面节省人工工作。在本文中,我们介绍了基于BERT的分类模型,以预测域和分类。我们还使用了来自经事实审查的文章的额外数据。我们利用额外的培训数据,为任务3A和任务3B实现了83.76%的宏观F1分,即85.55%的F1分。