Large Language Models (LLMs) are reshaping unsupervised learning by offering an unprecedented ability to perform text clustering based on their deep semantic understanding. However, their direct application is fundamentally limited by a lack of stateful memory for iterative refinement and the difficulty of managing cluster granularity. As a result, existing methods often rely on complex pipelines with external modules, sacrificing a truly end-to-end approach. We introduce LLM-MemCluster, a novel framework that reconceptualizes clustering as a fully LLM-native task. It leverages a Dynamic Memory to instill state awareness and a Dual-Prompt Strategy to enable the model to reason about and determine the number of clusters. Evaluated on several benchmark datasets, our tuning-free framework significantly and consistently outperforms strong baselines. LLM-MemCluster presents an effective, interpretable, and truly end-to-end paradigm for LLM-based text clustering.
翻译:大型语言模型(LLMs)凭借其深层的语义理解能力,正在重塑无监督学习领域,展现出前所未有的文本聚类性能。然而,其直接应用受到根本性限制:缺乏用于迭代优化的状态记忆,以及难以有效管理聚类粒度。因此,现有方法通常依赖包含外部模块的复杂流程,牺牲了真正的端到端处理方式。本文提出LLM-MemCluster,一种新颖的框架,将聚类重新构想为完全由LLM原生执行的任务。该框架利用动态记忆(Dynamic Memory)来注入状态感知能力,并通过双提示策略(Dual-Prompt Strategy)使模型能够推理并确定聚类数量。在多个基准数据集上的评估表明,我们的免调优框架显著且持续地超越了现有强基线方法。LLM-MemCluster为基于LLM的文本聚类提供了一种高效、可解释且真正端到端的范式。