Radio Access Network (RAN) slicing enables multiple logical networks to exist on top of the same physical infrastructure by allocating resources to distinct service groups, where radio resource scheduling plays a key role in ensuring compliance with slice-specific Service-Level Agreements (SLAs). Existing configuration-based or intent-driven Reinforcement Learning (RL) approaches usually rely on static mappings and SLA conversions. The current literature does not integrate natural language understanding with coordinated decision-making. To address these limitations, we propose an Agentic AI framework for 6G RAN slicing, driven by a super agent built using Hierarchical Decision Mamba (HDM) controllers and a Large Language Model (LLM). The super agent interprets operator intents and translates them into actionable goals using the LLM, which are used by HDM to coordinate inter-slice, intra-slice, and self-healing agents. Compared to transformer-based and reward-driven baselines, the proposed Agentic AI framework demonstrates consistent improvements across key performance indicators, including higher throughput, improved cell-edge performance, and reduced latency across different slices.
翻译:无线接入网(RAN)切片通过将资源分配给不同的服务组,实现在同一物理基础设施上承载多个逻辑网络,其中无线资源调度对于确保满足切片特定的服务等级协议(SLA)起着关键作用。现有的基于配置或意图驱动的强化学习(RL)方法通常依赖于静态映射和SLA转换。当前文献尚未将自然语言理解与协同决策相结合。为应对这些局限,我们提出一种用于6G RAN切片的智能体AI框架,其核心是由分层决策Mamba(HDM)控制器与大型语言模型(LLM)构建的超级智能体。该超级智能体通过LLM解析运营商意图并将其转化为可执行目标,随后由HDM协调切片间、切片内及自愈智能体。与基于Transformer和奖励驱动的基线方法相比,所提出的智能体AI框架在关键性能指标上均展现出持续改进,包括更高的吞吐量、更优的小区边缘性能以及跨不同切片的延迟降低。