Chinese medical question-answer matching is more challenging than the open-domain question answer matching in English. Even though the deep learning method has performed well in improving the performance of question answer matching, these methods only focus on the semantic information inside sentences, while ignoring the semantic association between questions and answers, thus resulting in performance deficits. In this paper, we design a series of interactive sentence representation learning models to tackle this problem. To better adapt to Chinese medical question-answer matching and take the advantages of different neural network structures, we propose the Crossed BERT network to extract the deep semantic information inside the sentence and the semantic association between question and answer, and then combine with the multi-scale CNNs network or BiGRU network to take the advantage of different structure of neural networks to learn more semantic features into the sentence representation. The experiments on the cMedQA V2.0 and cMedQA V1.0 dataset show that our model significantly outperforms all the existing state-of-the-art models of Chinese medical question answer matching.
翻译:中国的医学问答匹配比英语的开放式问答匹配更具挑战性。 尽管深层次的学习方法在改进问答匹配的性能方面表现良好, 但这些方法仅侧重于句子中的语义信息, 而忽略了问答之间的语义联系, 从而导致性能缺陷。 在本文中, 我们设计了一系列互动的句子代表学习模型来解决这个问题。 为了更好地适应中国医学问答匹配, 并且利用不同神经网络结构的优势, 我们建议交叉 BERT 网络在句子中提取深层次的语义信息, 以及问答之间的语义联系, 然后与多级CNN 网络或 BIGRU 网络相结合, 以利用神经网络的不同结构来学习句子表达中的语义特征。 有关 cMedQA V2. 0 和 cMedQA V1.0 数据集的实验显示, 我们的模型大大超越了所有中国医学问题解答的现有状态模式。