Deep neural networks have been shown to be very powerful modeling tools for many supervised learning tasks involving complex input patterns. However, they can also easily overfit to training set biases and label noises. In addition to various regularizers, example reweighting algorithms are popular solutions to these problems, but they require careful tuning of additional hyperparameters, such as example mining schedules and regularization hyperparameters. In contrast to past reweighting methods, which typically consist of functions of the cost value of each example, in this work we propose a novel meta-learning algorithm that learns to assign weights to training examples based on their gradient directions. To determine the example weights, our method performs a meta gradient descent step on the current mini-batch example weights (which are initialized from zero) to minimize the loss on a clean unbiased validation set. Our proposed method can be easily implemented on any type of deep network, does not require any additional hyperparameter tuning, and achieves impressive performance on class imbalance and corrupted label problems where only a small amount of clean validation data is available.
翻译:深神经网络被证明是许多涉及复杂输入模式的监管学习任务非常强大的模型工具,然而,它们也很容易被过度适应于培训设置的偏差和标签噪音。除了各种正规化者外,重加权算法是解决这些问题的流行办法,但是它们需要仔细调整额外的超参数,例如采矿时间表和使超参数正规化。与过去通常由每个实例的成本价值功能构成的重新加权方法相比,我们在此工作中提出了一种新的元学习算法,它学会根据梯度方向为培训范例分配权重。为了确定示例权重,我们的方法在目前的微型批次示例权重(从零开始)上实施一个元梯度梯度梯度梯度梯度梯度梯度梯度梯度梯度梯度梯度梯度梯度梯度梯度阶梯度梯度梯度梯度梯度梯度阶梯度阶梯度阶梯度阶梯度阶梯度阶梯度阶梯度阶梯度步骤,以尽量减少损失。我们提出的方法可以很容易在任何深度网络中实施,不需要任何额外的超参数调,在只有少量清洁验证数据的地方取得令人印象深刻的功能。