Intensive Care in-hospital mortality prediction has various clinical applications. Neural prediction models, especially when capitalising on clinical notes, have been put forward as improvement on currently existing models. However, to be acceptable these models should be performant and transparent. This work studies different attention mechanisms for clinical neural prediction models in terms of their discrimination and calibration. Specifically, we investigate sparse attention as an alternative to dense attention weights in the task of in-hospital mortality prediction from clinical notes. We evaluate the attention mechanisms based on: i) local self-attention over words in a sentence, and ii) global self-attention with a transformer architecture across sentences. We demonstrate that the sparse mechanism approach outperforms the dense one for the local self-attention in terms of predictive performance with a publicly available dataset, and puts higher attention to prespecified relevant directive words. The performance at the sentence level, however, deteriorates as sentences including the influential directive words tend to be dropped all together.
翻译:医院内密集护理死亡率预测具有各种临床应用。神经预测模型,特别是在利用临床注释时,已经作为现有模型的改进而提出来。然而,这些模型应当具有可接受性和透明度。这项工作研究临床神经预测模型在歧视和校准方面的不同关注机制。具体地说,我们从临床注释中调查在医院内死亡率预测任务中作为密集关注权重的替代物的注意力稀少问题。我们评估了关注机制,其依据是:(一) 当地对一个句子中的单词的自觉意识,和(二) 全球对一个变异器结构的自觉意识,跨句子。我们证明,稀疏机制方法在以公开的数据集预测性能方面优于密集的局部自我意识,并更加注意预先指定相关的指令词。然而,在句子上的表现恶化,因为包括有影响力的指令词在内的判决往往一起被删除。