Multi-source Domain Generalization (DG) measures a classifier's ability to generalize to new distributions of data it was not trained on, given several training domains. While several multi-source DG methods have been proposed, they incur additional complexity during training by using domain labels. Recent work has shown that a well-tuned Empirical Risk Minimization (ERM) training procedure, that is simply minimizing the empirical risk on the source domains, can outperform most existing DG methods. We identify several key candidate techniques to further improve ERM performance, such as better utilization of training data, model parameter selection, and weight-space regularization. We call the resulting method ERM++, and show it significantly improves the performance of DG on five multi-source datasets by over 5% compared to standard ERM, and beats state-of-the-art despite being less computationally expensive. Additionally, we demonstrate the efficacy of ERM++ on the WILDS-FMOW dataset, a challenging DG benchmark. We hope that ERM++ becomes a strong baseline for future DG research. Code is released at https://github.com/piotr-teterwak/erm_plusplus.
翻译:多源领域泛化(DG)衡量分类器在给定多个训练领域的情况下,对于其未经训练的新数据分布的泛化能力。尽管已经提出了几种多源DG方法,但它们在使用领域标签时会增加额外的复杂性。最新的研究表明,一种良好调整的经验风险最小化(ERM)训练程序,即在源领域上简单地最小化经验风险,可以超过大多数现有的DG方法。我们确定了几种关键的候选技术,以进一步提高ERM性能,例如更好地利用训练数据、模型参数选择和权重空间正则化。我们称之为ERM++并展示它在五个多源数据集上显著提高了DG性能,比标准ERM提高了5%以上,并超越了最先进的方法,同时计算代价更少。此外,我们还展示了ERM++在WILDS-FMOW数据集上的有效性,这是一个具有挑战性的DG基准。我们希望ERM++成为未来DG研究的强大基线。代码发布在https://github.com/piotr-teterwak/erm_plusplus。