Modern deep neural networks can easily overfit to biased training data containing corrupted labels or class imbalance. Sample re-weighting methods are popularly used to alleviate this data bias issue. Most current methods, however, require to manually pre-specify the weighting schemes as well as their additional hyper-parameters relying on the characteristics of the investigated problem and training data. This makes them fairly hard to be generally applied in practical scenarios, due to their significant complexities and inter-class variations of data bias situations. To address this issue, we propose a meta-model capable of adaptively learning an explicit weighting scheme directly from data. Specifically, by seeing each training class as a separate learning task, our method aims to extract an explicit weighting function with sample loss and task/class feature as input, and sample weight as output, expecting to impose adaptively varying weighting schemes to different sample classes based on their own intrinsic bias characteristics. Synthetic and real data experiments substantiate the capability of our method on achieving proper weighting schemes in various data bias cases, like the class imbalance, feature-independent and dependent label noise scenarios, and more complicated bias scenarios beyond conventional cases. Besides, the task-transferability of the learned weighting scheme is also substantiated, by readily deploying the weighting function learned on relatively smaller-scale CIFAR-10 dataset on much larger-scale full WebVision dataset. A performance gain can be readily achieved compared with previous SOAT ones without additional hyper-parameter tuning and meta gradient descent step. The general availability of our method for multiple robust deep learning issues, including partial-label learning, semi-supervised learning and selective classification, has also been validated.
翻译:现代深层神经网络可以轻易地超越含有腐败标签或阶级不平衡的有偏见的培训数据; 抽样重新加权方法被普遍用来缓解数据偏差问题; 然而,大多数现行方法要求根据所调查问题和培训数据的特点,手工预先确定加权办法及其额外的超参数; 这使得它们很难在实际情况下被普遍应用,因为其复杂性很大,数据偏差情况存在不同类别的差异。 为了解决这一问题,我们提议了一个能够适应性地直接从数据中学习一个明确的加权办法的元模型。 具体地说,通过将每个培训类视为一个不同的深度学习任务,我们的方法旨在提取一个明确的加权功能,以抽样损失和任务/类特性作为投入,以及作为产出的抽样重量,同时根据不同抽样的内在偏差特点,对不同的抽样类别实施适应性不同的加权办法。 合成和真实的数据实验证实了我们方法在各种数据偏差案例中实现适当加权办法的能力,如等级分类不平衡、特征依赖性和依赖性标签的噪声方案,以及更复杂的偏差假设情景,具体地将每个培训课期视为一种较严格的深的深度学习任务/级加权,此外,通过相对的逐步地学习SAL- 学习SAL-RO-RO-RO-RO-SL-SAL-SL-SL-S-SB-SAL-SL-SAL-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-