Early diagnosis of Alzheimer's disease (AD) is crucial in facilitating preventive care and to delay further progression. Speech based automatic AD screening systems provide a non-intrusive and more scalable alternative to other clinical screening techniques. Textual embedding features produced by pre-trained language models (PLMs) such as BERT are widely used in such systems. However, PLM domain fine-tuning is commonly based on the masked word or sentence prediction costs that are inconsistent with the back-end AD detection task. To this end, this paper investigates the use of prompt-based fine-tuning of PLMs that consistently uses AD classification errors as the training objective function. Disfluency features based on hesitation or pause filler token frequencies are further incorporated into prompt phrases during PLM fine-tuning. The exploit of the complementarity between BERT or RoBERTa based PLMs that are either prompt learning fine-tuned, or optimized using conventional masked word or sentence prediction costs, decision voting based system combination between them is further applied. Mean, standard deviation and the maximum among accuracy scores over 15 experiment runs are adopted as performance measurements for the AD detection system. Mean detection accuracy of 84.20% (with std 2.09%, best 87.5%) and 82.64% (with std 4.0%, best 89.58%) were obtained using manual and ASR speech transcripts respectively on the ADReSS20 test set consisting of 48 elderly speakers.
翻译:对阿尔茨海默氏病的早期诊断(AD)对于促进预防性护理和进一步推进至关重要。基于发言的自动自动自动筛选系统为其他临床筛查技术提供了一种非侵入性的、更可伸缩的替代方法。BERT等经过训练的语言模型产生的文字嵌入功能在这类系统中广泛使用。然而,PLM域微调通常基于与后端AD检测任务不相符的蒙面词或句数预测成本。为此,本文件调查了对始终使用AD分类错误作为培训目标功能的PLM的快速微调的使用。在PLM微调期间,基于犹豫或暂停信号信号频率的偏差功能被进一步纳入迅速的词组。BERT或RoBERTA的PLMs之间的互补性,要么是迅速学习,要么是使用传统的隐面词或句数预测成本,要么是进一步应用基于决定投票的系统组合。在15次实验中,标准偏差和最高精度分数是作为ADS检测系统的绩效测量标准,A.0.09%和A.009%的ADRR的精确度分别为8.5%和A.