Generalizing to new populations and domains in machine learning is still an open problem which has seen increased interest recently. In particular, clinical models show a significant performance drop when tested in settings not seen during training, e.g., new hospitals or population demographics. Recently proposed models for domain generalisation promise to alleviate this problem by learning invariant characteristics across environments, however, there is still scepticism about whether they improve over traditional training. In this work, we take a principled approach to identifying Out of Distribution (OoD) environments, motivated by the problem of cross-hospital generalization in critical care. We propose model-based and heuristic approaches to identify OoD environments and systematically compare models with different levels of held-out information. In particular, based on the assumption that models with access to OoD data should outperform other models, we train models across a range of experimental setups that include leave-one-hospital-out training and cross-sectional feature splits. We find that access to OoD data does not translate to increased performance, pointing to inherent limitations in defining potential OoD environments in the eICU Database potentially due to data harmonisation and sampling. Echoing similar results with other popular clinical benchmarks in the literature, new approaches are required to evaluate robust models in critical care.
翻译:特别是,临床模型显示,在培训期间没有看到的环境,例如新医院或人口统计中,在测试时业绩显著下降。最近提出的领域概括模型通过学习环境差异性特征来缓解这一问题,然而,人们仍然怀疑这些模型是否比传统培训有所改进。在这项工作中,我们采取原则性办法,查明关键护理跨医院普及问题引起的分配(OoD)环境。我们提出了基于模型和超强的方法,以确定OOOD环境,系统地将模型与不同水平的搁置信息进行比较。特别是,基于以下假设,即获取OOOD数据的模式应优于其他模式,我们培训一系列实验模型,其中包括请假一住院培训和跨部门特征分离。我们发现,获取OOD数据不能转化为更高绩效,我们发现在界定电子用户数据库中潜在的OOD环境方面存在内在的局限性,在确定电子用户数据库中潜在的潜在潜在潜在潜在环境,与不同水平的搁置信息比较。 特别是,基于这样一种假设,即获取OOD数据模型的模型的模型应比其他模型更为完善,因此,在临床处理中,还可能需要采用其他关键处理方法。