Reading comprehension (RC) has been studied in a variety of datasets with the boosted performance brought by deep neural networks. However, the generalization capability of these models across different domains remains unclear. To alleviate this issue, we are going to investigate unsupervised domain adaptation on RC, wherein a model is trained on labeled source domain and to be applied to the target domain with only unlabeled samples. We first show that even with the powerful BERT contextual representation, the performance is still unsatisfactory when the model trained on one dataset is directly applied to another target dataset. To solve this, we provide a novel conditional adversarial self-training method (CASe). Specifically, our approach leverages a BERT model fine-tuned on the source dataset along with the confidence filtering to generate reliable pseudo-labeled samples in the target domain for self-training. On the other hand, it further reduces domain distribution discrepancy through conditional adversarial learning across domains. Extensive experiments show our approach achieves comparable accuracy to supervised models on multiple large-scale benchmark datasets.
翻译:由于深神经网络提高了性能,对各种数据集进行了阅读理解(RC)的研究。然而,这些模型在不同领域的通用能力仍然不清楚。为了缓解这一问题,我们将调查RC上未经监督的域适应情况,该模型在标签源域上受过培训,并将应用于目标域,只有未贴标签的样本。我们首先显示,即使有了强大的BERT背景代表,当一个数据集培训的模型直接应用到另一个目标数据集时,其性能仍然不尽人意。为了解决这个问题,我们提供了一种新的有条件的对抗性自我培训方法(CASe)。具体地说,我们的方法利用了在源数据集上经过微调的BERT模型以及信任过滤器,在目标域生成可靠的假标签样本进行自我培训。另一方面,它通过跨域的有条件对抗性学习进一步缩小了区域分布差异。广泛的实验表明,我们的方法在多个大型基准数据集上与监督模型的精确度相当。