Transfer learning has been widely used in natural language processing through deep pretrained language models, such as Bidirectional Encoder Representations from Transformers and Universal Sentence Encoder. Despite the great success, language models get overfitted when applied to small datasets and are prone to forgetting when fine-tuned with a classifier. To remedy this problem of forgetting in transferring deep pretrained language models from one domain to another domain, existing efforts explore fine-tuning methods to forget less. We propose DeepEmotex an effective sequential transfer learning method to detect emotion in text. To avoid forgetting problem, the fine-tuning step is instrumented by a large amount of emotion-labeled data collected from Twitter. We conduct an experimental study using both curated Twitter data sets and benchmark data sets. DeepEmotex models achieve over 91% accuracy for multi-class emotion classification on test dataset. We evaluate the performance of the fine-tuned DeepEmotex models in classifying emotion in EmoInt and Stimulus benchmark datasets. The models correctly classify emotion in 73% of the instances in the benchmark datasets. The proposed DeepEmotex-BERT model outperforms Bi-LSTM result on the benchmark datasets by 23%. We also study the effect of the size of the fine-tuning dataset on the accuracy of our models. Our evaluation results show that fine-tuning with a large set of emotion-labeled data improves both the robustness and effectiveness of the resulting target task model.
翻译:在自然语言处理过程中,通过深层预先培训的语言模型,例如变换器和通用句号编码器的双向编码器演示模型,在自然语言处理中广泛使用转移学习。尽管取得了巨大成功,但语言模型在应用到小数据集时被过度装配,在与分类器进行微调时很容易忘记。为纠正在将深预先培训的语言模型从一个领域转移到另一个领域时忘记的问题,现有努力探索微调方法,以减少遗忘。我们建议深电motex是一种有效的连续传输学习方法,以探测文字中的情感。为了避免忘记问题,微调步骤用从Twitter收集的大量情感标签数据作工具。我们利用经调整的Twitter数据集和基准数据集进行实验性研究,我们利用经调整的TeepEmotex模型进行一项实验性研究,在测试数据集的多级情感分类方面实现超过91%的准确度。我们评估了在EmoInt和Stimulus基准数据集中对情感进行分类的精度分析的绩效。模型正确地将情感分解了73%的精度,我们在基准模型模型模型中收集的精确度数据定义的精确性数据调整结果。