Machine learning methods can be unreliable when deployed in domains that differ from the domains on which they were trained. To address this, we may wish to learn representations of data that are domain-invariant in the sense that we preserve data structure that is stable across domains, but throw out spuriously-varying parts. There are many representation-learning approaches of this type, including methods based on data augmentation, distributional invariances, and risk invariance. Unfortunately, when faced with any particular real-world domain shift, it is unclear which, if any, of these methods might be expected to work. The purpose of this paper is to show how the different methods relate to each other, and clarify the real-world circumstances under which each is expected to succeed. The key tool is a new notion of domain shift relying on the idea that causal relationships are invariant, but non-causal relationships (e.g., due to confounding) may vary.
翻译:机器学习方法在与培训领域不同的领域部署时可能不可靠。 为了解决这个问题,我们不妨学习对域变量的数据的表述方式,即我们维护跨域稳定的数据结构,但丢弃假相异的部分。这种类型的代表学习方法有许多,包括基于数据增强、分布差异和风险不易的方法。不幸的是,在遇到任何特定的现实世界域变换时,这些方法中哪些可能(如果有的话)可以发挥作用。本文的目的是说明不同方法如何相互关联,并澄清每个方法都有望成功的现实世界环境。关键工具是一个新的域变换概念,其依据的理念是因果关系是变化性的,但非因果关系(例如,由于粘结)可能有所不同。