Radio access network (RAN) technologies continue to witness massive growth, with Open RAN gaining the most recent momentum. In the O-RAN specifications, the RAN intelligent controller (RIC) serves as an automation host. This article introduces principles for machine learning (ML), in particular, reinforcement learning (RL) relevant for the O-RAN stack. Furthermore, we review state-of-the-art research in wireless networks and cast it onto the RAN framework and the hierarchy of the O-RAN architecture. We provide a taxonomy of the challenges faced by ML/RL models throughout the development life-cycle: from the system specification to production deployment (data acquisition, model design, testing and management, etc.). To address the challenges, we integrate a set of existing MLOps principles with unique characteristics when RL agents are considered. This paper discusses a systematic life-cycle model development, testing and validation pipeline, termed: RLOps. We discuss all fundamental parts of RLOps, which include: model specification, development and distillation, production environment serving, operations monitoring, safety/security and data engineering platform. Based on these principles, we propose the best practices for RLOps to achieve an automated and reproducible model development process.
翻译:在O-RAN规格中,RAN智能控制器(RIC)是一个自动化主机主机,它介绍了机器学习的原则(ML),尤其是与O-RAN堆叠有关的强化学习原则(RL),此外,我们审查了无线网络的最新研究,并将其纳入RAN框架和O-RAN结构的等级;我们从系统规格、开发和蒸馏、生产环境服务、操作监测、安全/安全和数据工程平台等发展周期到生产周期,对ML/RL/RL模型面临的挑战进行了分类。为了应对挑战,我们结合了一套在考虑RL代理时具有独特特点的现有MLOP原则。本文讨论了一个系统的生命周期模型开发、测试和验证管道,称为:RLOPs。我们讨论了RLOps的所有基本组成部分,其中包括:示范规格、开发和蒸馏、生产环境服务、操作监测、安全/安全和数据工程平台。我们根据这些原则,建议一个自动化的RLOPS发展过程。