The domain generalization (DG) problem setting challenges a model trained on multiple known data distributions to generalise well on unseen data distributions. Due to its practical importance, a large number of methods have been proposed to address this challenge. However much of the work in general purpose DG is heuristically motivated, as the DG problem is hard to model formally; and recent evaluations have cast doubt on existing methods' practical efficacy -- in particular compared to a well tuned empirical risk minimisation baseline. We present a novel learning-theoretic generalisation bound for DG that bounds unseen domain generalisation performance in terms of the model's Rademacher complexity. Based on this, we conjecture that existing methods' efficacy or lack thereof is largely determined by an empirical risk vs predictor complexity trade-off, and demonstrate that their performance variability can be explained in these terms. Algorithmically, this analysis suggests that domain generalisation should be achieved by simply performing regularised ERM with a leave-one-domain-out cross-validation objective. Empirical results on the DomainBed benchmark corroborate this.
翻译:一般化(DG)问题挑战了在多种已知数据分配方面受过培训的模型,以便很好地概括各种无形数据分布。由于其实际重要性,提出了大量方法来应对这一挑战。然而,由于DG问题难以正式建模,一般用途DG的许多工作具有超自然动机,因为DG问题很难正式建模;最近的评价使人们对现有方法的实际效力产生怀疑 -- -- 特别是相对于一个经过良好调整的实验风险最小化基线而言。我们为DG提出了一种新的学习理论概括化,约束了DG在模型Rademacher复杂性方面的无形域概括性业绩。基于这一点,我们推测,现有方法的有效性或缺乏效力主要取决于经验风险与预测性复杂性交易,并表明其绩效的可变性可以用这些术语加以解释。从理论上讲,这一分析表明,应该通过仅仅实施正规化的、有请假的跨度验证目标来实现领域的一般化。在DGGG中,根据模型的 Rademacher复杂度,我们推测,现有方法的有效性或缺乏效力主要取决于经验性风险与预测性交易。