Following the pandemic outbreak, several works have proposed to diagnose COVID-19 with deep learning in computed tomography (CT); reporting performance on-par with experts. However, models trained/tested on the same in-distribution data may rely on the inherent data biases for successful prediction, failing to generalize on out-of-distribution samples or CT with different scanning protocols. Early attempts have partly addressed bias-mitigation and generalization through augmentation or re-sampling, but are still limited by collection costs and the difficulty of quantifying bias in medical images. In this work, we propose Mixing-AdaSIN; a bias mitigation method that uses a generative model to generate de-biased images by mixing texture information between different labeled CT scans with semantically similar features. Here, we use Adaptive Structural Instance Normalization (AdaSIN) to enhance de-biasing generation quality and guarantee structural consistency. Following, a classifier trained with the generated images learns to correctly predict the label without bias and generalizes better. To demonstrate the efficacy of our method, we construct a biased COVID-19 vs. bacterial pneumonia dataset based on CT protocols and compare with existing state-of-the-art de-biasing methods. Our experiments show that classifiers trained with de-biased generated images report improved in-distribution performance and generalization on an external COVID-19 dataset.
翻译:疫情爆发后,一些工作提议对COVID-19进行诊断,在计算断层成像术(CT)中进行深层学习;与专家一起报告业绩;然而,根据同样的分配数据进行培训/测试的模型可能依赖内在数据偏差来成功预测,未能对分发之外的样本或CT采用不同的扫描规程进行概括化;早期尝试部分地解决了通过增强或再取样来减少偏差和概括化的问题,但仍然受到收集成本和医疗图像偏见量化困难的限制;在这项工作中,我们提议混合-AdaSIN;一种偏差缓解方法,利用一种基因模型,将带有类似特征的不同贴标签的CT扫描之间的纹理信息混在一起,产生不偏差的图像。在这里,我们使用适应性结构常态常态化(AdaSIN),通过增强生成的生成质量或再抽样,保证结构一致性。随后,接受过培训的D级分类人员学会正确预测标签,而没有偏差和概括化。为了展示我们的方法的功效,我们用有偏差的COVID-19-19级平流图像对比的外部测试方法,用我们经过训练的实验室的模型展示了一种状态。