Empirical Risk Minimization (ERM) based machine learning algorithms have suffered from weak generalization performance on data obtained from out-of-distribution (OOD). To address this problem, Invariant Risk Minimization (IRM) objective was suggested to find invariant optimal predictor which is less affected by the changes in data distribution. However, even with such progress, IRMv1, the practical formulation of IRM, still shows performance degradation when there are not enough training data, and even fails to generalize to OOD, if the number of spurious correlations is larger than the number of environments. In this paper, to address such problems, we propose a novel meta-learning based approach for IRM. In this method, we do not assume the linearity of classifier for the ease of optimization, and solve ideal bi-level IRM objective with Model-Agnostic Meta-Learning (MAML) framework. Our method is more robust to the data with spurious correlations and can provide an invariant optimal classifier even when data from each distribution are scarce. In experiments, we demonstrate that our algorithm not only has better OOD generalization performance than IRMv1 and all IRM variants, but also addresses the weakness of IRMv1 with improved stability.
翻译:风险最小化(ERM)基于机床学习算法(ERM)基于机床的机床学习算法(ERM)在从分配外获得的数据上,普遍化表现不力,因而受到影响。为解决这一问题,我们建议了不变化风险最小化(IRM)目标,以寻找不变化的最佳预测器,这种预测器受数据分布变化的影响较小。然而,即使取得了这样的进步,IRMv1, IRM1的实用配方,在没有足够的培训数据的情况下,仍然显示性能退化,甚至未能对OOD进行概括化,如果虚假相关关系的数量大于环境的数量。在本文件中,为了解决这些问题,我们提出了一种新的基于元学习的IMM方法。在这种方法中,我们并不假定分类器的直线性,以方便优化,解决理想的双级IRM目标与模型- Agnoict met-Lain(MAM) 框架相比,我们的方法对具有虚假相关性的数据更为有力,即使每次分发的数据很少,也能够提供不变化的最佳分类。在试验中,我们证明我们的演算法不仅比IMM1所有性变式的系统更稳定,而且也比IMMMRMRM1改进了所有的版本。