Deep Reinforcement Learning for Optimization of RAN Slicing Relying on Control- and User-Plane Separation.

IEEE Internet Things J.(2024)

引用 0|浏览1
暂无评分
摘要
The rapid development of radio access network (RAN) slicing and control-and user-plane separation (CUPS) has created a new paradigm for future networks, namely CUPS-based RAN slicing. In this paper, we formulate the utility optimization problems of the CUPS-based RAN slicing system and propose a Lyapunov-based deep reinforcement learning (L-DRL) framework to solve them. Specifically, we propose that the CP and UP slices should control their respective power and subcarrier resources. Firstly, we provide coverage-driven slices in the CP for coverage control and data-driven slices in the UP for diverse user requests, where we consider the influence of coverage-driven slices on data-driven slices. Secondly, we define the system’s utilities as income minus cost, and we formulate the utility maximization problem of the UP as a mixed-integer nonlinear programming problem (MINLP), which is NP-hard because it considers both continuous actions (densities deployment and power allocation) and discrete action (subcarrier allocation). Furthermore, we design an alternating optimization method for the CP and UP based on the densities of deployment. Finally, we develop a novel L-DRL framework for mixed-action optimization problems and propose a specific Lyapunov-based asynchronous advantage actor-critic (L-A3C) algorithm. Simulation results demonstrate that our proposed L-A3C algorithm outperforms the standard A3C algorithm in terms of the convergence while achieving higher performance than Lyapunov optimization. Moreover, our proposed CUPS-based RAN slicing scheme surpasses the benchmark RAN slicing schemes in terms of the achievable rate and delay.
更多
查看译文
关键词
Radio access network slicing,control-and User-plane separation,Lyapunov optimization,A3C
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要