In recent years, there has been a resurgence of interest in non-autoregressive text generation in the context of general language modeling. Unlike the well-established autoregressive language modeling paradigm, which has a plethora of standard training and inference libraries, implementations of non-autoregressive language modeling have largely been bespoke making it difficult to perform systematic comparisons of different methods. Moreover, each non-autoregressive language model typically requires it own data collation, loss, and prediction logic, making it challenging to reuse common components. In this work, we present the XLM python package, which is designed to make implementing small non-autoregressive language models faster with a secondary goal of providing a suite of small pre-trained models (through a companion xlm-models package) that can be used by the research community. The code is available at https://github.com/dhruvdcoder/xlm-core.
翻译:近年来,在通用语言建模背景下,非自回归文本生成的研究兴趣再度兴起。与已建立完善标准训练和推理库的自回归语言建模范式不同,非自回归语言建模的实现大多为定制化方案,导致难以对不同方法进行系统性比较。此外,每种非自回归语言模型通常需要独立的数据整理、损失函数和预测逻辑,这使得通用组件的复用面临挑战。本研究提出XLM Python软件包,其核心设计目标是加速小型非自回归语言模型的实现,次要目标是通过配套的xlm-models软件包为研究社区提供一套可用于实验的预训练小型模型。代码发布于https://github.com/dhruvdcoder/xlm-core。