In this paper, we have worked on interpretability, trust, and understanding of the decisions made by models in the form of classification tasks. The task is divided into 3 subtasks. The first task consists of determining Binary Sexism Detection. The second task describes the Category of Sexism. The third task describes a more Fine-grained Category of Sexism. Our work explores solving these tasks as a classification problem by fine-tuning transformer-based architecture. We have performed several experiments with our architecture, including combining multiple transformers, using domain adaptive pretraining on the unlabelled dataset provided by Reddit and Gab, Joint learning, and taking different layers of transformers as input to a classification head. Our system (with team name Attention) was able to achieve a macro F1 score of 0.839 for task A, 0.5835 macro F1 score for task B and 0.3356 macro F1 score for task C at the Codalab SemEval Competition. Later we improved the accuracy of Task B to 0.6228 and Task C to 0.3693 in the test set.
翻译:在本文中,我们研究了模型的分类任务中对决策的可解释性、可信度和理解性。该任务分为3个子任务。第一个任务是确定二元性别歧视检测。第二个任务描述了性别歧视的类别。第三个任务描述了更细致的性别歧视类别。我们的工作探索将这些任务作为分类问题通过微调基于transformer的架构来解决。我们对我们的体系结构进行了多次实验,包括组合多个transformer,使用Reddit和Gab提供的未标记数据集上的领域自适应预训练,联合学习以及将transformer的不同层作为分类头的输入。我们的系统(团队名Attention)在 Codalab SemEval比赛中为任务A达到了0.839的宏F1得分,任务B的宏F1得分为0.5835,任务C的宏F1得分为0.3356。后来我们在测试集中将任务B的准确度提高到了0.6228,任务C的准确度提高到了0.3693。