Much of the existing linguistic data in many languages of the world is locked away in non-digitized books and documents. Optical character recognition (OCR) can be used to produce digitized text, and previous work has demonstrated the utility of neural post-correction methods that improve the results of general-purpose OCR systems on recognition of less-well-resourced languages. However, these methods rely on manually curated post-correction data, which are relatively scarce compared to the non-annotated raw images that need to be digitized. In this paper, we present a semi-supervised learning method that makes it possible to utilize these raw images to improve performance, specifically through the use of self-training, a technique where a model is iteratively trained on its own outputs. In addition, to enforce consistency in the recognized vocabulary, we introduce a lexically-aware decoding method that augments the neural post-correction model with a count-based language model constructed from the recognized texts, implemented using weighted finite-state automata (WFSA) for efficient and effective decoding. Results on four endangered languages demonstrate the utility of the proposed method, with relative error reductions of 15-29%, where we find the combination of self-training and lexically-aware decoding essential for achieving consistent improvements. Data and code are available at https://shrutirij.github.io/ocr-el/.
翻译:光字符识别(OCR)可用于制作数字化文本,先前的工作已经证明神经后校正方法的效用,这些方法改进了一般用途OCR系统在承认资源较少的语言方面的结果;然而,这些方法依赖人工整理后校正数据,与需要数字化的非附加说明的原始图像相比,这些数据相对较少。本文介绍了一种半监督的学习方法,它使得有可能利用这些原始图像改进性能,特别是通过使用自我培训,一种模型在自身产出方面迭接培训的技术。此外,为了在公认的词汇中加强一致性,我们采用了一种具有法律意识的校正方法,用根据公认的文本建立的基于数字的校正后语言模型来增强神经后校正模型,采用加权的定数自动解码(WFSA)来高效和有效地解译。关于四种濒危语言的成果展示了以自己为主的图像的实用性,特别是使用自我培训,在15-29级的常规中,我们找到了一致的自译/自译法。