Transformer-based models for transfer learning have the potential to achieve high prediction accuracies on text-based supervised learning tasks with relatively few training data instances. These models are thus likely to benefit social scientists that seek to have as accurate as possible text-based measures but only have limited resources for annotating training data. To enable social scientists to leverage these potential benefits for their research, this paper explains how these methods work, why they might be advantageous, and what their limitations are. Additionally, three Transformer-based models for transfer learning, BERT (Devlin et al. 2019), RoBERTa (Liu et al. 2019), and the Longformer (Beltagy et al. 2020), are compared to conventional machine learning algorithms on three applications. Across all evaluated tasks, textual styles, and training data set sizes, the conventional models are consistently outperformed by transfer learning with Transformers, thereby demonstrating the benefits these models can bring to text-based social science research.
翻译:以变换器为基础的转让学习模型有可能在基于文本的监督下学习任务上实现高预测值,而培训数据实例相对较少,因此这些模型有可能使社会科学家受益,他们寻求尽可能精确的基于文本的措施,但用于说明培训数据的资源有限。为了使社会科学家能够利用这些潜在的好处来进行研究,本文件解释了这些方法如何发挥作用,为什么它们可能有利,以及它们的局限性。此外,三个基于变换器的转让学习模型,即BERT(Devlin等人,2019年)、RoBERTA(Liu等人,2019年)和Longexu(Beltaty等人,2020年)都与三种应用的传统机器学习算法进行了比较。在所有被评估的任务、文字风格和培训数据集大小中,传统模型始终比与变换器的学习要好,从而展示这些模型能够给基于文本的社会科学研究带来的益处。