The use of transfer learning methods is largely responsible for the present breakthrough in Natural Learning Processing (NLP) tasks across multiple domains. In order to solve the problem of sentiment detection, we examined the performance of four different types of well-known state-of-the-art transformer models for text classification. Models such as Bidirectional Encoder Representations from Transformers (BERT), Robustly Optimized BERT Pre-training Approach (RoBERTa), a distilled version of BERT (DistilBERT), and a large bidirectional neural network architecture (XLNet) were proposed. The performance of the four models that were used to detect disaster in the text was compared. All the models performed well enough, indicating that transformer-based models are suitable for the detection of disaster in text. The RoBERTa transformer model performs best on the test dataset with a score of 82.6% and is highly recommended for quality predictions. Furthermore, we discovered that the learning algorithms' performance was influenced by the pre-processing techniques, the nature of words in the vocabulary, unbalanced labeling, and the model parameters.
翻译:转让学习方法的使用是目前自然学习处理(NLP)跨多个领域任务的突破的主要原因。为了解决情绪检测问题,我们研究了四种不同类型已知的最新变压器模型的性能,以进行文本分类。模型,如变压器的双向电解码显示器(BERT),强力优化的BERT预培训方法(ROBERTA),一个蒸馏版的BERT(DutilBERT)和一个大型双向神经网络结构(XLNet),已经提出。在文本中用于探测灾害的四种模型的性能得到了比较。所有模型都表现良好,表明基于变压器的模型都适合在文本中探测灾害。RoBERTA变压器模型在测试数据集中最出色地表现了82.6%的分数,并被高度推荐用于质量预测。此外,我们发现学习算法的性能受到预处理技术、词汇的性质、不平衡的标签和模型参数的影响。</s>