This paper investigates the effectiveness of sentence-level transformers for zero-shot offensive span identification on a code-mixed Tamil dataset. More specifically, we evaluate rationale extraction methods of Local Interpretable Model Agnostic Explanations (LIME) \cite{DBLP:conf/kdd/Ribeiro0G16} and Integrated Gradients (IG) \cite{DBLP:conf/icml/SundararajanTY17} for adapting transformer based offensive language classification models for zero-shot offensive span identification. To this end, we find that LIME and IG show baseline $F_{1}$ of 26.35\% and 44.83\%, respectively. Besides, we study the effect of data set size and training process on the overall accuracy of span identification. As a result, we find both LIME and IG to show significant improvement with Masked Data Augmentation and Multilabel Training, with $F_{1}$ of 50.23\% and 47.38\% respectively. \textit{Disclaimer : This paper contains examples that may be considered profane, vulgar, or offensive. The examples do not represent the views of the authors or their employers/graduate schools towards any person(s), group(s), practice(s), or entity/entities. Instead they are used to emphasize only the linguistic research challenges.}
翻译:本文调查了在泰米尔语代码混合数据集中用于零射进攻性跨度识别的句级变压器的有效性。 更具体地说, 我们评估了当地解释性示范解释(LIME)\cite{DBLP:conf/ kddd/Ribeiro0G16}和综合梯度(IG)\cite{DBLP:conf/icml/Sundaranjanty17}和综合梯度变压器变压器变压器攻击性语言分类模型用于零射射线跨度识别的效果。 为此, 我们发现LIME和IG分别展示了26.35美元1美元和44.83美元的基准值。 此外, 我们研究了数据集大小和培训过程对确定跨度总体准确性的影响。 结果,我们发现LIME和IG都展示了使用保护性数据放大和多标签培训的显著改进之处, 分别以50.23美元( ⁇ ) 和47.38 。\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\