Classic machine learning methods are built on the $i.i.d.$ assumption that training and testing data are independent and identically distributed. However, in real scenarios, the $i.i.d.$ assumption can hardly be satisfied, rendering the sharp drop of classic machine learning algorithms' performances under distributional shifts, which indicates the significance of investigating the Out-of-Distribution generalization problem. Out-of-Distribution (OOD) generalization problem addresses the challenging setting where the testing distribution is unknown and different from the training. This paper serves as the first effort to systematically and comprehensively discuss the OOD generalization problem, from the definition, methodology, evaluation to the implications and future directions. Firstly, we provide the formal definition of the OOD generalization problem. Secondly, existing methods are categorized into three parts based on their positions in the whole learning pipeline, namely unsupervised representation learning, supervised model learning and optimization, and typical methods for each category are discussed in detail. We then demonstrate the theoretical connections of different categories, and introduce the commonly used datasets and evaluation metrics. Finally, we summarize the whole literature and raise some future directions for OOD generalization problem. The summary of OOD generalization methods reviewed in this survey can be found at http://out-of-distribution-generalization.com.
翻译:典型的机器学习方法基于“$i.i.d.d.”这一假设,即培训和测试数据是独立的,而且分布完全相同。然而,在实际假设中,“$i.i.i.d.d.”这一假设几乎无法满足,使得经典机器学习算法在分布式转换中的性能急剧下降,这表明了调查 " 分配外普遍化 " 问题的重要性。 " 分配外分配(OOD)一般化问题 " (OOOD)一般化问题解决了测试分布不为人知和不同于培训的具有挑战性的环境。本文是系统、全面地讨论OOD一般化问题,从定义、方法、评价到影响和未来方向的首次努力。首先,我们提供了OOOD一般化问题的正式定义。第二,现有方法根据其在整个学习过程中的立场分为三个部分,即未经监督的代表学习、监督的示范学习和优化以及每个类别的典型方法。我们随后展示了不同类别的理论联系,并介绍了常用的数据集和评估指标。最后,我们总结了OOD一般化问题的整个文献,并提出未来方向。