The memorization effect of deep learning hinders its performance to effectively generalize on test set when learning with noisy labels. Prior study has discovered that epistemic uncertainty techniques are robust when trained with noisy labels compared with neural networks without uncertainty estimation. They obtain prolonged memorization effect and better generalization performance under the adversarial setting of noisy labels. Due to its superior performance amongst other selected epistemic uncertainty methods under noisy labels, we focus on Monte Carlo Dropout (MCDropout) and investigate why it is robust when trained with noisy labels. Through empirical studies on datasets MNIST, CIFAR-10, Animal-10n, we deep dive into three aspects of MCDropout under noisy label setting: 1. efficacy: understanding the learning behavior and test accuracy of MCDropout when training set contains artificially generated or naturally embedded label noise; 2. representation volatility: studying the responsiveness of neurons by examining the mean and standard deviation on each neuron's activation; 3. network sparsity: investigating the network support of MCDropout in comparison with deterministic neural networks. Our findings suggest that MCDropout further sparsifies and regularizes the deterministic neural networks and thus provides higher robustness against noisy labels.
翻译:深层学习的记忆化效应阻碍其表现,使其无法在使用噪音标签学习时,在测试测试集成时有效普及; 先前的研究发现,与神经网络相比,在不作不确定性估计的情况下,在与神经网络相比的吵动标签中,经培养的标签过于吵闹的标签,在训练噪音标签时,隐性不确定性技术具有很强性能; 在吵闹标签的对抗性设置下,它们获得长期的记忆化效应和更好的概括性性表现; 由于它优于在噪音标签下其他选定的隐性不确定性方法,我们把重点放在蒙特卡洛脱落(MCDropout)上,并调查它为什么在接受噪音标签培训时很强; 通过对MNIST、CIFAR-10、动物-10n进行实验性研究,我们深潜入MCDropout的三个方面:1. 功效:当训练集成含有人工生成或自然嵌入的标签噪音时,了解MCDroproput的学习行为和测试性能; 2. 代表性波动:通过检查每个神经活的平均值和标准偏差来研究神经活性; 3 网络紧张性:调查MCDropoputout 与确定性神经网络相比,对网络的支持。 我们的调查结论认为MCD-ROpropmentsut 进一步调整了高度网络。