Understanding the internal physiological changes accompanying the aging process is an important aspect of medical image interpretation, with the expected changes acting as a baseline when reporting abnormal findings. Deep learning has recently been demonstrated to allow the accurate estimation of patient age from chest X-rays, and shows potential as a health indicator and mortality predictor. In this paper we present a novel comparative study of the relative performance of radiologists versus state-of-the-art deep learning models on two tasks: (a) patient age estimation from a single chest X-ray, and (b) ranking of two time-separated images of the same patient by age. We train our models with a heterogeneous database of 1.8M chest X-rays with ground truth patient ages and investigate the limitations on model accuracy imposed by limited training data and image resolution, and demonstrate generalisation performance on public data. To explore the large performance gap between the models and humans on these age-prediction tasks compared with other radiological reporting tasks seen in the literature, we incorporate our age prediction model into a conditional Generative Adversarial Network (cGAN) allowing visualisation of the semantic features identified by the prediction model as significant to age prediction, comparing the identified features with those relied on by clinicians.
翻译:了解随着老龄化过程而来的内在生理变化是医学形象解释的一个重要方面,预期的变化在报告异常结果时可以作为基准。最近已经展示了深度学习,以便从胸前X光中准确估计病人年龄,并显示出作为健康指标和死亡率预测者的潜力。在本文中,我们介绍了对放射科医生相对性能的比较研究,与最先进的深层次学习模式的比较研究,这两项任务是:(a) 从单一胸部X光中估计病人年龄,和(b) 按年龄对同一病人的两种时间分离的图像进行分级。我们用一个1.8M胸X光的混合数据库对模型进行了培训,该数据库具有地面真相病人年龄,调查有限的培训数据和图像分辨率对模型准确性的限制,并展示公共数据的总体性表现。为了探索这些模型与人类在年龄定位任务上与文献中看到的其他辐射报告任务之间的巨大性能差距,我们将我们的年龄预测模型纳入一个有条件的Generational Adversarial网络(cAN),以便能够对预测模型所查明的精度特征进行可视化,从而将诊断模型所根据的临床特征进行重大年龄预测。