Automated simplification models aim to make input texts more readable. Such methods have the potential to make complex information accessible to a wider audience, e.g., providing access to recent medical literature which might otherwise be impenetrable for a lay reader. However, such models risk introducing errors into automatically simplified texts, for instance by inserting statements unsupported by the corresponding original text, or by omitting key information. Providing more readable but inaccurate versions of texts may in many cases be worse than providing no such access at all. The problem of factual accuracy (and the lack thereof) has received heightened attention in the context of summarization models, but the factuality of automatically simplified texts has not been investigated. We introduce a taxonomy of errors that we use to analyze both references drawn from standard simplification datasets and state-of-the-art model outputs. We find that errors often appear in both that are not captured by existing evaluation metrics, motivating a need for research into ensuring the factual accuracy of automated simplification models.
翻译:自动化简化模式旨在使输入文本更易于读取,这些方法有可能使更广大的受众能够查阅复杂的信息,例如,提供最近医学文献的检索,否则非普通读者可能无法查阅,然而,这些模式有可能在自动简化文本中引入错误,例如插入未经相应原始文本支持的语句,或省略关键信息。提供更易读但不准确的文本,在许多情况下可能比不提供这种访问更糟糕。在综合化模式中,事实准确性(以及缺乏准确性)的问题受到高度关注,但自动简化文本的实际情况尚未调查。我们采用了一种错误分类法,用以分析从标准简化数据集和最新模型产出中提取的参考文献。我们发现,现有评价指标没有记录到的错误经常出现在这两种错误中,因此有必要进行研究,以确保自动简化模型的实际准确性。