Very large language models (LLMs), such as GPT-3 and Codex have achieved state-of-the-art performance on several natural-language tasks, and show great promise also for code. A particularly exciting aspect of LLMs is their knack for few-shot and zero-shot learning: they can learn to perform a task with very few examples. Few-shotting has particular synergies in software engineering, where there are a lot of phenomena (identifier names, APIs, terminology, coding patterns) that are known to be highly project-specific. However, project-specific data can be quite limited, especially early in the history of a project; thus the few-shot learning capacity of LLMs might be very relevant. In this paper, we investigate the use few-shot training with the very large GPT (Generative Pre-trained Transformer) Codex model, and find evidence suggesting that one can significantly surpass state-of-the-art models for code-summarization, leveraging project-specific training.
翻译:大型语言模型(LLMs),如GPT-3和Codx等大型语言模型(LLMs)在一些自然语言任务上取得了最先进的表现,也表现出了对代码的巨大希望。LLMs的一个特别令人振奋的方面是他们能够进行几分和零分学习的技巧:他们可以用很少的例子来学习执行任务。很少的射击在软件工程方面特别具有协同作用,因为已知有许多现象(识别名称、API、术语、编码模式)与项目非常具体化。然而,具体项目的数据可能非常有限,特别是在一个项目的早期;因此LLMs的微小的学习能力可能非常相关。在本文中,我们调查了与非常庞大的GPT(受过训练的预变型)代码模型进行的几分手培训,并发现有证据表明,在代码汇总、利用具体项目培训方面,有一个可能大大超过最先进的模式。