Automatic diagnosis has attracted increasing attention but remains challenging due to multi-step reasoning. Recent works usually address it by reinforcement learning methods. However, these methods show low efficiency and require taskspecific reward functions. Considering the conversation between doctor and patient allows doctors to probe for symptoms and make diagnoses, the diagnosis process can be naturally seen as the generation of a sequence including symptoms and diagnoses. Inspired by this, we reformulate automatic diagnosis as a symptoms Sequence Generation (SG) task and propose a simple but effective automatic Diagnosis model based on Transformer (Diaformer). We firstly design the symptom attention framework to learn the generation of symptom inquiry and the disease diagnosis. To alleviate the discrepancy between sequential generation and disorder of implicit symptoms, we further design three orderless training mechanisms. Experiments on three public datasets show that our model outperforms baselines on disease diagnosis by 1%, 6% and 11.5% with the highest training efficiency. Detailed analysis on symptom inquiry prediction demonstrates that the potential of applying symptoms sequence generation for automatic diagnosis.
翻译:自动诊断已引起越来越多的关注,但由于多步推理,仍然具有挑战性。最近的工作通常通过强化学习方法加以解决。然而,这些方法显示低效率,需要特定任务的报酬功能。考虑到医生和病人之间的对话允许医生对症状进行检测和诊断,诊断过程自然可以被视为包括症状和诊断在内的序列的生成。受此启发,我们重新将自动诊断作为症状序列生成(SG)的任务,并提议一个简单而有效的基于变异器(Diaexer)的自动诊断模型。我们首先设计症状关注框架,以学习症状调查和疾病诊断的生成。为了减轻连续生成与隐含症状的紊乱之间的差异,我们进一步设计了三个无序的培训机制。对三个公共数据集的实验表明,我们的模型比疾病诊断基线高出1%、6%和11.5%,而培训效率最高。对症状调查预测的详细分析表明,将症状序列生成用于自动诊断的可能性。