Digital Twin Enabled Q-Learning for Flying Base Station Placement: Impact of Varying Environment and Model Errors

ISORC(2023)

引用 0|浏览0
暂无评分
摘要
This paper considers a use case of flying base station placement enabled by digital twin (DT), and demonstrates how DT can help reduce the impact of a non-stationary environment on reinforcement learning (RL). RL is able to learn the optimal policy via interacting with a specific environment. However, it has been observed that RL is very sensitive to environment change, mainly because the environment variation disturbs RL training/learning. A possible approach is to execute the RL process in the DT using snapshots of the environment (parameters), and update the parameters at a proper frequency. The DT-RL bundled approach takes advantage of computing resources in the DT, speeds up the process and saves battery energy of the flying base station, and more importantly, mitigates the non-stationary impact on RL. Specifically, the use case is about quickly connecting mobile users with an aerial bass station. The base station is autonomously and optimally placed according to a predefined criterion to connect scattered slow-movement users. Q-learning, a common type of RL, is employed as a solution to the optimization of base station placement. Tailored for this application, a two-stage base station placement algorithm is proposed and evaluated. For the configuration considered in this paper, numerical results suggest that 1) the Q-learning algorithm run solely in the physical space does not work due to intolerable time consuming and optimization divergence, and 2) the proposed scheme can catch up with random slow movement of mobile users, and tolerate certain measurement and model errors. With necessary modification and extension, the proposed framework could be applied to other DT-assisted cyber-physical systems.
更多
查看译文
关键词
Aerial/flying base station,digital twin (DT),reinforcement learning (RL),Q-learning
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要