Real-world large-scale medical image analysis (MIA) datasets have three challenges: 1) they contain noisy-labelled samples that affect training convergence and generalisation, 2) they usually have an imbalanced distribution of samples per class, and 3) they normally comprise a multi-label problem, where samples can have multiple diagnoses. Current approaches are commonly trained to solve a subset of those problems, but we are unaware of methods that address the three problems simultaneously. In this paper, we propose a new training module called Non-Volatile Unbiased Memory (NVUM), which non-volatility stores running average of model logits for a new regularization loss on noisy multi-label problem. We further unbias the classification prediction in NVUM update for imbalanced learning problem. We run extensive experiments to evaluate NVUM on new benchmarks proposed by this paper, where training is performed on noisy multi-label imbalanced chest X-ray (CXR) training sets, formed by Chest-Xray14 and CheXpert, and the testing is performed on the clean multi-label CXR datasets OpenI and PadChest. Our method outperforms previous state-of-the-art CXR classifiers and previous methods that can deal with noisy labels on all evaluations. Our code is available at https://github.com/FBLADL/NVUM.
翻译:现实世界大规模医学图像分析(MIA)数据集有三个挑战:(1) 含有影响培训趋同和总体化的杂乱标签样本,(2) 通常每类样本分布不均,(3) 通常由多标签问题组成,样品可以进行多重诊断。 目前的方法通常经过培训,可以解决其中的一组问题,但我们不知道同时解决这三个问题的方法。 在本文件中,我们提议一个新的培训模块,名为“无流动性无偏见内存(NVUUM)”,该模块的非挥发性商店平均使用模型登录表来计算噪音多标签多标签问题的新正规化损失。我们进一步在NVUM更新时对不平衡学习问题进行分类预测,而样本则由多个标签有多重标签不平衡的胸部X-射线(CXR)和CheXpert(NUM)之间新的基准进行评估。我们的方法是用之前的CX-RR和PadchestRexx(ANDR)之间的标准评估。