Due to domain shift, machine learning systems typically fail to generalize well to domains different from those of training data, which is what domain generalization (DG) aims to address. Although various DG methods have been developed, most of them lack interpretability and require domain labels that are not available in many real-world scenarios. This paper presents a novel DG method, called HMOE: Hypernetwork-based Mixture of Experts (MoE), which does not rely on domain labels and is more interpretable. MoE proves effective in identifying heterogeneous patterns in data. For the DG problem, heterogeneity arises exactly from domain shift. HMOE uses hypernetworks taking vectors as input to generate experts' weights, which allows experts to share useful meta-knowledge and enables exploring experts' similarities in a low-dimensional vector space. We compare HMOE with other DG algorithms under a fair and unified benchmark-DomainBed. Our extensive experiments show that HMOE can divide mixed-domain data into distinct clusters that are surprisingly more consistent with human intuition than original domain labels. Compared to other DG methods, HMOE shows competitive performance and achieves SOTA results in some cases.
翻译:由于域变,机器学习系统通常无法将不同领域与培训数据领域(即领域通用(DG)的目的在于解决的领域)相容。虽然已经开发了各种DG方法,但大多数DG方法缺乏可解释性,需要在许多现实世界情景中不具备的域名标签。本文介绍了一种新型DG方法,称为HMOE:基于超网络的专家混合体(MOE),它不依赖域名标签,更易解释。MOE证明在确定数据差异模式方面是有效的。对于DG问题,异质性完全产生于域变换。HMOE使用超额网络作为输入来生成专家的权重,让专家分享有用的元知识,并探索专家在低维矢量空间中的相似性。我们在公平和统一的基准-DomainBe下将HMOE与其他DG算法进行比较。我们的广泛实验显示,HMOE可以将混合数据分成不同组群,这与人类直觉比原始域标更为一致。与SOTA相比,HMOE显示一些具有竞争力的成绩和SOTA案例取得的结果。</s>