Existing federated learning models that follow the standard risk minimization paradigm of machine learning often fail to generalize in the presence of spurious correlations in the training data. In many real-world distributed settings, spurious correlations exist due to biases and data sampling issues on distributed devices or clients that can erroneously influence models. Current generalization approaches are designed for centralized training and attempt to identify features that have an invariant causal relationship with the target, thereby reducing the effect of spurious features. However, such invariant risk minimization approaches rely on apriori knowledge of training data distributions which is hard to obtain in many applications. In this work, we present a generalizable federated learning framework called FedGen, which allows clients to identify and distinguish between spurious and invariant features in a collaborative manner without prior knowledge of training distributions. We evaluate our approach on real-world datasets from different domains and show that FedGen results in models that achieve significantly better generalization than current federated learning approaches.
翻译:遵循标准风险最小化机器学习模式的现有联邦学习模式往往无法在培训数据存在虚假关联的情况下推广。在许多现实世界分布的环境中,由于分布式设备或客户的偏差和数据抽样问题,可能会错误影响模型,存在虚假关联。目前的概括化方法是为了集中培训,并试图查明与目标有不变因果关系的特征,从而减少虚假特征的影响。然而,这种变化式风险最小化方法依赖于培训数据分布的优先知识,而在许多应用中很难获得这种知识。在这项工作中,我们提出了一个通用的联邦化学习框架,即FedGen,使客户能够在没有事先培训分布知识的情况下,以协作方式识别和区分虚假和不易变特征。我们评估了我们在不同领域对真实世界数据集的做法,并表明FedGen在模型中取得了比目前联合学习方法更好的普及性。