Mental disorders represent a critical global health challenge, and social media is increasingly viewed as a vital resource for real-time digital phenotyping and intervention. Large Language Models (LLMs) offer stronger semantic understanding and reasoning than traditional deep learning, but their use in high-stakes clinical settings is limited by hallucinations and the lack of persistent memory. However, existing literature has not sufficiently investigated how advanced enhancement techniques, specifically Retrieval-Augmented Generation (RAG) and Agentic systems, can address these reliability and reasoning limitations. Here, we systematically survey the evolving landscape of LLM-based methods for social media mental disorder analysis, spanning standard pretrained language models, RAG to mitigate hallucinations and contextual gaps, and agentic systems for autonomous reasoning and multi-step intervention. We organize existing work by technical paradigm and clinical target, extending beyond common internalizing disorders to include psychotic disorders and externalizing behaviors. Additionally, the paper comprehensively evaluates the performance of LLMs, including the impact of RAG, across various tasks. This work establishes a unified benchmark for the field, paving the way for the development of trustworthy, autonomous AI systems that can deliver precise and explainable mental health support.
翻译:精神障碍是一项严峻的全球性健康挑战,而社交媒体日益被视为实时数字表型分析和干预的重要资源。大语言模型(LLMs)相比传统深度学习方法具备更强的语义理解与推理能力,但其在高风险临床场景中的应用仍受限于幻觉问题及缺乏持久记忆。然而,现有文献尚未充分探讨先进的增强技术——特别是检索增强生成(RAG)与智能体系统——如何应对这些可靠性与推理方面的局限。本文系统性地综述了基于LLM的社交媒体精神障碍分析方法的演进脉络,涵盖标准预训练语言模型、用于缓解幻觉与上下文缺失的RAG技术,以及实现自主推理与多步骤干预的智能体系统。我们依据技术范式与临床目标对现有研究进行分类梳理,其范围不仅包括常见的内化障碍,亦扩展至精神病性障碍与外化行为。此外,本文全面评估了LLMs(包括RAG技术的影响)在多项任务中的性能表现。本研究为该领域建立了统一的基准,为开发可信赖、自主且能提供精准、可解释心理健康支持的人工智能系统铺平了道路。