Deep learning (DL) models are state-of-the-art in segmenting anatomical and disease regions of interest (ROIs) in medical images. Particularly, a large number of DL-based techniques have been reported using chest X-rays (CXRs). However, these models are reportedly trained on reduced image resolutions for reasons related to the lack of computational resources. Literature is sparse in discussing the optimal image resolution to train these models for segmenting the Tuberculosis (TB)-consistent lesions in CXRs. In this study, we investigated the performance variations using an Inception-V3 UNet model using various image resolutions with/without lung ROI cropping and aspect ratio adjustments, and (ii) identified the optimal image resolution through extensive empirical evaluations to improve TB-consistent lesion segmentation performance. We used the Shenzhen CXR dataset for the study which includes 326 normal patients and 336 TB patients. We proposed a combinatorial approach consisting of storing model snapshots, optimizing segmentation threshold and test-time augmentation (TTA), and averaging the snapshot predictions, to further improve performance with the optimal resolution. Our experimental results demonstrate that higher image resolutions are not always necessary, however, identifying the optimal image resolution is critical to achieving superior performance.
翻译:深度学习模式(DL)是医学图像中引起关注的解剖和疾病区域分解的最先进方法,特别是,报告了大量基于DL的技术,使用胸前X光(CXRs)报告了大量基于DL的技术,然而,据报告,由于缺乏计算资源,这些模型受到降低图像分辨率的培训。在讨论这些模型的最佳图像分辨率以在CXRs中分离肺结核(TB)同质损伤时,文献稀少。在本研究中,我们使用一种“感知V3UNet”模型调查性能差异,使用各种图像分辨率,使用肺部ROI作物和侧面比调整,(二) 通过广泛的实证评估,确定最佳图像分辨率。我们利用Shenzen CXR数据集进行研究,其中包括326个正常病人和336个肺结核病人。我们建议采用一种组合式方法,包括存储模型快照、优化分解阈值和测试时间增强(TTTTA),以及将快速预测进行,以便进一步改进性能,然而最佳性能显示最佳性能。