Neural models, with their ability to provide novel representations, have shown promising results in prediction tasks in healthcare. However, patient demographics, medical technology, and quality of care change over time. This often leads to drop in the performance of neural models for prospective patients, especially in terms of their calibration. The deep kernel learning (DKL) framework may be robust to such changes as it combines neural models with Gaussian processes, which are aware of prediction uncertainty. Our hypothesis is that out-of-distribution test points will result in probabilities closer to the global mean and hence prevent overconfident predictions. This in turn, we hypothesise, will result in better calibration on prospective data. This paper investigates DKL's behaviour when facing a temporal shift, which was naturally introduced when an information system that feeds a cohort database was changed. We compare DKL's performance to that of a neural baseline based on recurrent neural networks. We show that DKL indeed produced superior calibrated predictions. We also confirm that the DKL's predictions were indeed less sharp. In addition, DKL's discrimination ability was even improved: its AUC was 0.746 (+- 0.014 std), compared to 0.739 (+- 0.028 std) for the baseline. The paper demonstrated the importance of including uncertainty in neural computing, especially for their prospective use.
翻译:具有提供新表现形式能力的内核模型(DKL)在健康护理的预测任务中显示出了令人乐观的结果。然而,耐心的人口统计、医疗技术和护理质量随着时间推移而发生变化。这往往导致未来患者神经模型性能下降,特别是校准。深内核学习(DKL)框架对于将神经模型和高萨进程(这些进程都意识到预测的不确定性)相结合的变化可能非常有力。我们的假设是,分配以外的测试点将产生更接近全球平均值的概率,从而防止过度的不自信预测。反过来,我们假设,这将导致未来数据的更精确校准。在面对时间变化时,DKL的行为会得到调查。当一个为组合数据库提供材料的信息系统被改变时,这一框架自然被引入。我们将DKL的性能与基于经常性神经网络的神经基线的性能表现进行比较。我们显示,DKL的测试点将产生更精确的校准预测。我们还证实,DKL的预测确实会降低未来数据的精确度。我们还证实,对于未来数据的校准性数据,特别是DKL的S-MAR+40的分析能力在模型中展示了它们的模型中的重要性。