Deep Neural Networks (DNNs) have been shown to be susceptible to memorization or overfitting in the presence of noisily labelled data. For the problem of robust learning under such noisy data, several algorithms have been proposed. A prominent class of algorithms rely on sample selection strategies, motivated by curriculum learning. For example, many algorithms use the `small loss trick' wherein a fraction of samples with loss values below a certain threshold are selected for training. These algorithms are sensitive to such thresholds, and it is difficult to fix or learn these thresholds. Often, these algorithms also require information such as label noise rates which are typically unavailable in practice. In this paper, we propose a data-dependent, adaptive sample selection strategy that relies only on batch statistics of a given mini-batch to provide robustness against label noise. The algorithm does not have any additional hyperparameters for sample selection, does not need any information on noise rates, and does not need access to separate data with clean labels. We empirically demonstrate the effectiveness of our algorithm on benchmark datasets.
翻译:深神经网络(DNNS)被证明很容易在有贴有新标签的数据的情况下进行记忆化或过度配置。对于在如此吵闹的数据下进行强力学习的问题,提出了几种算法。一个突出的算法类别依赖于抽样选择战略,其动机是学习课程。例如,许多算法使用“小损失技巧”,其中选择了一小部分损失值低于某一阈值的样本来进行培训。这些算法对此类阈值敏感,难以固定或学习这些阈值。这些算法通常也需要一些信息,如标签噪音率等,而实际上通常无法使用。在本文件中,我们提出了一个依赖数据、适应性抽样选择战略,仅依靠某一微型批量的批次统计数据来提供稳健的标签噪音。该算法不需要任何额外的超参数来选择样品,不需要任何关于噪音率的信息,也不需要使用与清洁标签分开的数据。我们从经验上展示了基准数据集的算法的有效性。