This paper is a submission to the Alzheimer's Dementia Recognition through Spontaneous Speech (ADReSS) challenge, which aims to develop methods that can assist in the automated prediction of severity of Alzheimer's Disease from speech data. We focus on acoustic and natural language features for cognitive impairment detection in spontaneous speech in the context of Alzheimer's Disease Diagnosis and the mini-mental state examination (MMSE) score prediction. We proposed a model that obtains unimodal decisions from different LSTMs, one for each modality of text and audio, and then combines them using a gating mechanism for the final prediction. We focused on sequential modelling of text and audio and investigated whether the disfluencies present in individuals' speech relate to the extent of their cognitive impairment. Our results show that the proposed classification and regression schemes obtain very promising results on both development and test sets. This suggests Alzheimer's Disease can be detected successfully with sequence modeling of the speech data of medical sessions.
翻译:本文是针对阿尔茨海默氏病通过自发演讲(ADRESS)确认老年痴呆症(DADRESS)的挑战提交的,其目的是制定一些方法,协助从语音数据自动预测阿尔茨海默氏病的严重性,我们侧重于在阿尔茨海默氏病诊断和小型状态测试(MMSE)得分预测背景下自发说话认知障碍时的声学和自然语言特征,我们提出了一个模型,从不同的LSTM中获取单式决定,每个文本和音频模式都有一个,然后使用最后预测的格子机制将它们结合起来。我们侧重于文字和音频的顺序建模,并调查个人言论中的不便是否与其认知障碍程度有关。我们的结果显示,拟议的分类和回归计划在发展和测试组合中都获得了非常有希望的结果。这表明,通过对医学会议语音数据进行顺序建模,可以成功地检测阿尔茨海默氏病。