Knowledge enhanced pre-trained language models (K-PLMs) are shown to be effective for many public tasks in the literature but few of them have been successfully applied in practice. To address this problem, we propose K-AID, a systematic approach that includes a low-cost knowledge acquisition process for acquiring domain knowledge, an effective knowledge infusion module for improving model performance, and a knowledge distillation component for reducing the model size and deploying K-PLMs on resource-restricted devices (e.g., CPU) for real-world application. Importantly, instead of capturing entity knowledge like the majority of existing K-PLMs, our approach captures relational knowledge, which contributes to better-improving sentence-level text classification and text matching tasks that play a key role in question answering (QA). We conducted a set of experiments on five text classification tasks and three text matching tasks from three domains, namely E-commerce, Government, and Film&TV, and performed online A/B tests in E-commerce. Experimental results show that our approach is able to achieve substantial improvement on sentence-level question answering tasks and bring beneficial business value in industrial settings.
翻译:为解决这一问题,我们建议K-AID, 这是一种系统化的方法,包括低成本知识获取过程,用于获取域知识,一个有效的知识注入模块,以及一个知识提炼部分,用于缩小模型规模和在资源限制装置(如CPU)上部署K-PLMs,用于实际应用。重要的是,我们的方法不是捕捉现有K-PLMs中的大多数这样的实体知识,而是捕捉关系知识,这有助于更好地改进判决一级的文本分类和文本匹配任务,这些任务在回答问题(QA)中起着关键作用。我们就5项文本分类任务和3项文本匹配任务进行了一系列实验,这3项任务来自3个领域,即电子商务、政府、电影和电视,并在电子商务中进行在线A/B测试。实验结果表明,我们的方法能够大大改进判决层面的回答任务,并在工业环境中带来有益的商业价值。