Mental health is a critical issue in modern society, and mental disorders could sometimes turn to suicidal ideation without adequate treatment. Early detection of mental disorders and suicidal ideation from social content provides a potential way for effective social intervention. Recent advances in pretrained contextualized language representations have promoted the development of several domain-specific pretrained models and facilitated several downstream applications. However, there are no existing pretrained language models for mental healthcare. This paper trains and release two pretrained masked language models, i.e., MentalBERT and MentalRoBERTa, to benefit machine learning for the mental healthcare research community. Besides, we evaluate our trained domain-specific models and several variants of pretrained language models on several mental disorder detection benchmarks and demonstrate that language representations pretrained in the target domain improve the performance of mental health detection tasks.
翻译:在现代社会,心理健康是一个关键问题,精神失常有时在没有适当治疗的情况下会转向自杀思想。从社会内容中早期发现精神失常和自杀思想为有效的社会干预提供了潜在途径。在经过培训的背景化语言表现方面最近取得的进展促进了若干特定领域预先培训的模式的开发,并为若干下游应用提供了便利。然而,目前还没有关于精神保健的预先培训语言模式。这一纸张培训和释放了两个经过培训的隐蔽语言模式,即心理康复者和精神失常者,以便为心理保健研究界提供机器学习。此外,我们评估了我们经过培训的特定领域模式以及若干关于精神失常检测基准的经过培训的语言模式的几种模式,并表明在目标领域经过培训的语言表现将改善心理健康检测任务的绩效。