The introduction of pretrained language models has reduced many complex task-specific NLP models to simple lightweight layers. An exception to this trend is coreference resolution, where a sophisticated task-specific model is appended to a pretrained transformer encoder. While highly effective, the model has a very large memory footprint -- primarily due to dynamically-constructed span and span-pair representations -- which hinders the processing of complete documents and the ability to train on multiple instances in a single batch. We introduce a lightweight end-to-end coreference model that removes the dependency on span representations, handcrafted features, and heuristics. Our model performs competitively with the current standard model, while being simpler and more efficient.
翻译:采用经过培训的语文模型将许多复杂的任务特定NLP模型减为简单的轻量级。 这一趋势的一个例外是共同参考分辨率,即将复杂的任务特定模型附在经过培训的变压器编码器编码器上。虽然该模型非常有效,但它的记忆足迹非常大 -- -- 主要是由于动态构造的跨度和横幅表示法 -- -- 这妨碍了完整文件的处理和单批多例培训的能力。我们引入了轻量端对端共同参考模型,消除了对跨度展示、手工艺特征和超感学的依赖性。我们的模型在与现行标准模型竞争的同时更简单、更有效率。