Contrastive pre-training on distant supervision has shown remarkable effectiveness in improving supervised relation extraction tasks. However, the existing methods ignore the intrinsic noise of distant supervision during the pre-training stage. In this paper, we propose a weighted contrastive learning method by leveraging the supervised data to estimate the reliability of pre-training instances and explicitly reduce the effect of noise. Experimental results on three supervised datasets demonstrate the advantages of our proposed weighted contrastive learning approach compared to two state-of-the-art non-weighted baselines.Our code and models are available at: https://github.com/YukinoWan/WCL
翻译:关于远程监督的对立培训前培训在改进监督关系提取任务方面显示了显著的实效,但是,现有方法忽视了培训前阶段远程监督的内在噪音,在本文中,我们建议采用加权对比学习方法,利用监督数据估计培训前情况的可靠性,明确减少噪音的影响。三个监督数据集的实验结果表明,与两个最先进的非加权基线相比,我们拟议的加权对比学习方法具有优势。我们的代码和模型可在以下网址查阅:https://github.com/YukinoWan/WCL。