Recent studies show that deep learning models achieve good performance on medical imaging tasks such as diagnosis prediction. Among the models, multimodality has been an emerging trend, integrating different forms of data such as chest X-ray (CXR) images and electronic medical records (EMRs). However, most existing methods incorporate them in a model-free manner, which lacks theoretical support and ignores the intrinsic relations between different data sources. To address this problem, we propose a knowledge-driven and data-driven framework for lung disease diagnosis. By incorporating domain knowledge, machine learning models can reduce the dependence on labeled data and improve interpretability. We formulate diagnosis rules according to authoritative clinical medicine guidelines and learn the weights of rules from text data. Finally, a multimodal fusion consisting of text and image data is designed to infer the marginal probability of lung disease. We conduct experiments on a real-world dataset collected from a hospital. The results show that the proposed method outperforms the state-of-the-art multimodal baselines in terms of accuracy and interpretability.
翻译:最近的研究显示,深层次学习模式在诊断预测等医学成像任务方面表现良好。在模型中,多式联运是一个新出现的趋势,整合了不同形式的数据,如胸部X光(CXR)图像和电子医疗记录(EMRs),然而,大多数现有方法都以没有模型的方式纳入这些模型,缺乏理论支持,忽视了不同数据来源之间的内在关系。为解决这一问题,我们提议了一个由知识驱动和数据驱动的肺病诊断框架。通过纳入领域知识,机器学习模式可以减少对标签数据的依赖,并改进可解释性。我们根据权威临床医学准则制定诊断规则,并从文本数据中学习规则的分量。最后,由文本和图像数据组成的多式联运组合旨在推断肺病的边际概率。我们在从医院收集的真实世界数据集上进行实验。结果显示,拟议的方法在准确性和可解释性方面超过了最先进的多式基线。