Identifying adverse and hostile content on the web and more particularly, on social media, has become a problem of paramount interest in recent years. With their ever increasing popularity, fine-tuning of pretrained Transformer-based encoder models with a classifier head are gradually becoming the new baseline for natural language classification tasks. In our work, we explore the gains attributed to Task Adaptive Pretraining (TAPT) prior to fine-tuning of Transformer-based architectures. We specifically study two problems, namely, (a) Coarse binary classification of Hindi Tweets into Hostile or Not, and (b) Fine-grained multi-label classification of Tweets into four categories: hate, fake, offensive, and defamation. Building up on an architecture which takes emojis and segmented hashtags into consideration for classification, we are able to experimentally showcase the performance upgrades due to TAPT. Our system (with team name 'iREL IIIT') ranked first in the 'Hostile Post Detection in Hindi' shared task with an F1 score of 97.16% for coarse-grained detection and a weighted F1 score of 62.96% for fine-grained multi-label classification on the provided blind test corpora.
翻译:我们在工作中探索了在对基于变压器的建筑进行微调之前的适应性预备训练(TAPT)的成果。我们特别研究了两个问题,即:(a) 粗糙的二进制分解成敌对或非敌对,以及(b) 精细的多标签分类Tweets,分为四种类别:仇恨、假、冒犯和诽谤。在建构一个结构时,将emoji和分解标签纳入分类考虑。我们可以实验性地展示TAPT的性能升级。我们的系统(团队名称“iREL IIIT”)在“印地语中Hid Tweets误差检测”中名列第一,F1分为97.16%,用于粗劣的检测,F1分为62.96的盲段标签,用于精密的多级的F1分。我们用F1分制的F1分,用于精密的检测,用于在标签上提供精密的F1-96%的F1分分级的F1分级。