In Natural Language Processing, the use of pre-trained language models has been shown to obtain state-of-the-art results in many downstream tasks such as sentiment analysis, author identification and others. In this work, we address the use of these methods for personality classification from text. Focusing on the Myers-Briggs (MBTI) personality model, we describe a series of experiments in which the well-known Bidirectional Encoder Representations from Transformers (BERT) model is fine-tuned to perform MBTI classification. Our main findings suggest that the current approach significantly outperforms well-known text classification models based on bag-of-words and static word embeddings alike across multiple evaluation scenarios, and generally outperforms previous work in the field.
翻译:在自然语言处理中,通过使用预先培训的语言模型,在许多下游任务中取得了最先进的成果,如情绪分析、作者识别等。在这项工作中,我们从文字中探讨使用这些方法进行个性分类。我们侧重于Myers-Briggs(MBTI)个性模型,我们描述了一系列实验,在实验中,众所周知的变形器双向编码显示模型(BERT)模型经过精细调整,以进行MBTI分类。我们的主要发现表明,目前的方法大大优于以一袋字和静态字嵌入多种评价情景为基础而广为人知的文本分类模型,并普遍优于以往的实地工作。