A key assumption in supervised learning is that training and test data follow the same probability distribution. However, this fundamental assumption is not always satisfied in practice, e.g., due to changing environments, sample selection bias, privacy concerns, or high labeling costs. Transfer learning (TL) relaxes this assumption and allows us to learn under distribution shift. Classical TL methods typically rely on importance-weighting -- a predictor is trained based on the training losses weighted according to the importance (i.e., the test-over-training density ratio). However, as real-world machine learning tasks are becoming increasingly complex, high-dimensional, and dynamical, novel approaches are explored to cope with such challenges recently. In this article, after introducing the foundation of TL based on importance-weighting, we review recent advances based on joint and dynamic importance-predictor estimation. Furthermore, we introduce a method of causal mechanism transfer that incorporates causal structure in TL. Finally, we discuss future perspectives of TL research.
翻译:监督学习的一个关键假设是,培训和测试数据的概率分布相同,然而,由于环境变化、抽样选择偏好、隐私问题或标签成本高等原因,这一基本假设在实践中并非总能得到满足。转让学习(TL)放松了这一假设,使我们能够在分配变化中学习。经典TL方法通常依赖重要性加权 -- -- 预测员根据按重要性加权计算的培训损失来培训(即测试-超培训密度比率)。然而,随着现实世界的机器学习任务日益复杂、高度和动态,我们最近探索了应对此类挑战的新办法。在本篇文章中,在引入基于重要性加权基础的TL基础之后,我们根据联合和动态重要性-指标估计来审查最近的进展。此外,我们引入了一种将因果结构纳入TL的因果机制转移方法。最后,我们讨论了TL研究的未来前景。