Recently, chest X-ray report generation, which aims to automatically generate descriptions of given chest X-ray images, has received growing research interests. The key challenge of chest X-ray report generation is to accurately capture and describe the abnormal regions. In most cases, the normal regions dominate the entire chest X-ray image, and the corresponding descriptions of these normal regions dominate the final report. Due to such data bias, learning-based models may fail to attend to abnormal regions. In this work, to effectively capture and describe abnormal regions, we propose the Contrastive Attention (CA) model. Instead of solely focusing on the current input image, the CA model compares the current input image with normal images to distill the contrastive information. The acquired contrastive information can better represent the visual features of abnormal regions. According to the experiments on the public IU-X-ray and MIMIC-CXR datasets, incorporating our CA into several existing models can boost their performance across most metrics. In addition, according to the analysis, the CA model can help existing models better attend to the abnormal regions and provide more accurate descriptions which are crucial for an interpretable diagnosis. Specifically, we achieve the state-of-the-art results on the two public datasets.
翻译:最近,旨在自动生成对特定胸部X光图像描述的胸X光报告的生成,引起了越来越多的研究兴趣;胸X光报告生成的主要挑战是准确捕捉和描述异常区域;在多数情况下,正常区域占整个胸X光图像的主导,而对这些正常区域的相应描述则占最后报告的主导地位;由于这些数据偏差,学习模型可能无法进入异常区域;在这项工作中,为了有效捕捉和描述异常区域,我们提议了抗争注意模式(CA)模式;CA模式不仅侧重于当前的输入图像,而且将目前的输入图像与正常图像进行比较,以提取对比性信息;获得的对比性信息可以更好地反映异常区域的视觉特征;根据公共IU-X光和MIMIC-CX数据集的实验,将我们的CA纳入几个现有模型,可以提高它们在大多数计量指标中的性能;此外,根据分析,CA模型可以帮助现有模型更好地处理异常区域,并提供对可解释性诊断至关重要的更准确的描述。具体地说,我们实现了两种公共数据结果。