It is expensive to evaluate the results of Machine Translation(MT), which usually requires manual translation as a reference. Machine Translation Quality Estimation (QE) is a task of predicting the quality of machine translations without relying on any reference. Recently, the emergence of predictor-estimator framework which trains the predictor as a feature extractor and estimator as a QE predictor, and pre-trained language models(PLM) have achieved promising QE performance. However, we argue that there are still gaps between the predictor and the estimator in both data quality and training objectives, which preclude QE models from benefiting from a large number of parallel corpora more directly. Based on previous related work that have alleviated gaps to some extent, we propose a novel framework that provides a more accurate direct pretraining for QE tasks. In this framework, a generator is trained to produce pseudo data that is closer to the real QE data, and a estimator is pretrained on these data with novel objectives that are the same as the QE task. Experiments on widely used benchmarks show that our proposed framework outperforms existing methods, without using any pretraining models such as BERT.
翻译:机械翻译通常需要人工翻译,而机器翻译通常需要人工翻译作为参考。 机器翻译质量估计(QE)是一项不依靠任何参考而预测机器翻译质量的任务。 最近,出现了一个预测器-估计器框架,将预测器培训成一个特征提取器和估计器,将预测器培训成一个特征提取器和估计器作为QE预测器,预先培训的语言模型(PLM)取得了有希望的量化评估性能。然而,我们认为,在数据质量和培训目标方面,预测器和估测器之间仍然存在着差距,这使得QE模型无法更直接地受益于大量平行的子公司。基于以前的相关工作,在某种程度上缩小了差距,我们提出了一个新的框架,为QE任务提供了更准确的直接培训。在这个框架内,对发电机进行了培训,以产生更接近真实的量化评估数据,对这些数据进行了预先培训,其新目标与量化评估仪任务相同。 广泛使用的基准实验显示,我们使用的任何框架都显示,没有采用任何现有方法,而没有培训。