Machine learning (ML) models are typically optimized for their accuracy on a given dataset. However, this predictive criterion rarely captures all desirable properties of a model, in particular how well it matches a domain expert's understanding of a task. Underspecification refers to the existence of multiple models that are indistinguishable in their in-domain accuracy, even though they differ in other desirable properties such as out-of-distribution (OOD) performance. Identifying these situations is critical for assessing the reliability of ML models. We formalize the concept of underspecification and propose a method to identify and partially address it. We train multiple models with an independence constraint that forces them to implement different functions. They discover predictive features that are otherwise ignored by standard empirical risk minimization (ERM), which we then distill into a global model with superior OOD performance. Importantly, we constrain the models to align with the data manifold to ensure that they discover meaningful features. We demonstrate the method on multiple datasets in computer vision (collages, WILDS-Camelyon17, GQA) and discuss general implications of underspecification. Most notably, in-domain performance cannot serve for OOD model selection without additional assumptions.
翻译:然而,这一预测性标准很少能捕捉模型的所有可取特性,特别是它与域专家对任务的理解相匹配的程度。具体程度低是指存在多种模型,这些模型在内部准确性方面是无法区分的,尽管这些模型在其他理想特性上存在差异,例如分配外的性能。查明这些情况对于评估ML模型的可靠性至关重要。我们正式确定具体程度不足的概念,并提议一种识别和部分解决的方法。我们培训了多种具有独立性限制的模型,迫使它们执行不同功能。它们发现了预测性特征,这些特性被标准的实验风险最小化(ERM)所忽略,然后我们将其注入一个具有高级OOD性能的全球模型。重要的是,我们限制这些模型与数据元相协调,以确保它们发现有意义的特性。我们演示了计算机视觉中多个数据集的方法(代码、WILDS-Camelyon17, GQA),并讨论了不作详细描述的一般模型影响。最明显的是,在OD选择中无法为额外提供。