While a large number of pre-trained models of source code have been successfully developed and applied to a variety of software engineering (SE) tasks in recent years, our understanding of these pre-trained models is arguably fairly limited. With the goal of advancing our understanding of these models, we perform the first systematic empirical comparison of 19 recently-developed pre-trained models of source code on 13 SE tasks. To gain additional insights into these models, we adopt a recently-developed 4-dimensional categorization of pre-trained models, and subsequently investigate whether there are correlations between different categories of pre-trained models and their performances on different SE tasks.
翻译:虽然近年来成功地开发了大量经过培训的源代码模型,并应用于各种软件工程任务,但我们对这些经过培训的模型的理解可以说是相当有限的,为了增进对这些模型的理解,我们对最近开发的19个经过培训的源代码模型进行了首次系统经验比较,涉及13个经培训的源代码模型的任务。为了对这些模型有更多的了解,我们采用了最近开发的对经过培训的模型的四维分类,随后调查了不同类别经过培训的模型及其在各种东南欧任务上的表现之间是否相互关联。