Natural language processing technology has rapidly improved automated grammatical error correction tasks, and the community begins to explore document-level revision as one of the next challenges. To go beyond sentence-level automated grammatical error correction to NLP-based document-level revision assistant, there are two major obstacles: (1) there are few public corpora with document-level revisions being annotated by professional editors, and (2) it is not feasible to elicit all possible references and evaluate the quality of revision with such references because there are infinite possibilities of revision. This paper tackles these challenges. First, we introduce a new document-revision corpus, TETRA, where professional editors revised academic papers sampled from the ACL anthology which contain few trivial grammatical errors that enable us to focus more on document- and paragraph-level edits such as coherence and consistency. Second, we explore reference-less and interpretable methods for meta-evaluation that can detect quality improvements by document revision. We show the uniqueness of TETRA compared with existing document revision corpora and demonstrate that a fine-tuned pre-trained language model can discriminate the quality of documents after revision even when the difference is subtle. This promising result will encourage the community to further explore automated document revision models and metrics in future.
翻译:自然语言处理技术迅速改进了自动语法错误校正任务,社区开始探索作为下一个挑战之一的文件校正,作为下一个挑战之一的文件级修改。为了超越以判决一级自动语法错误校正为主的NLP文件级修订助理,存在两大障碍:(1) 很少有公共社团,专业编辑对文件级的修订加注;(2) 由于有无限的修订可能性,因此不可能收集所有可能的参考资料,并用这些参考资料来评估修订的质量。本文件应对这些挑战。首先,我们引入一个新的文件审校材料(TETRA),其中专业编辑从ACL炭文学中抽取的经校文修改的学术论文,其中几乎没有小的语法错误,使我们能够更多地关注文件级和段落级的编辑,如一致性和一致性。第二,我们探索无参考和可解释的元评价方法,通过文件修订可以检测质量的改进。我们展示TETRA与现有文件审校校校公司的独特性,并展示经过微调整的经培训的语文模式,在修订文件后,甚至会在审校订文件质量上进一步探索标准。