Understanding generalization is crucial to confidently engineer and deploy machine learning models, especially when deployment implies a shift in the data domain. For such domain adaptation problems, we seek generalization bounds which are tractably computable and tight. If these desiderata can be reached, the bounds can serve as guarantees for adequate performance in deployment. However, in applications where deep neural networks are the models of choice, deriving results which fulfill these remains an unresolved challenge; most existing bounds are either vacuous or has non-estimable terms, even in favorable conditions. In this work, we evaluate existing bounds from the literature with potential to satisfy our desiderata on domain adaptation image classification tasks, where deep neural networks are preferred. We find that all bounds are vacuous and that sample generalization terms account for much of the observed looseness, especially when these terms interact with measures of domain shift. To overcome this and arrive at the tightest possible results, we combine each bound with recent data-dependent PAC-Bayes analysis, greatly improving the guarantees. We find that, when domain overlap can be assumed, a simple importance weighting extension of previous work provides the tightest estimable bound. Finally, we study which terms dominate the bounds and identify possible directions for further improvement.
翻译:理解一般化对于自信地设计和部署机器学习模式至关重要,特别是在部署意味着数据领域变化的情况下。对于这类领域适应问题,我们寻求能够容易地兼容和紧凑的概括化界限。如果能够达到这些偏差范围,那么这些界限可以作为在部署中适当表现的保障。然而,在深神经网络是选择模式的应用程序中,实现这些选择的结果仍然是一个尚未解决的挑战;大多数现有界限要么是空的,要么是非可贵的术语,甚至在有利的条件下也是如此。在这项工作中,我们评估文献的现有界限,这些界限有可能满足我们对领域适应图像分类任务的偏差。我们发现,所有界限都是空洞的,样本化术语占了观察到的松散程度的大部分,特别是当这些术语与域变换措施相互作用时。要克服这一点并取得最接近的结果,我们把每个界限与最近的数据依赖PAC-Bayes分析结合起来,大大改进了保障。我们发现,在假设域重叠时,可以假设一个简单的重要范围,即缩小了我们先前工作的主要方向。</s>