Approximate Markov Perfect Equilibrium of Joint Offloading Policy for Multi-IV Using Reward-Shared Distributed Method

IEEE Transactions on Intelligent Vehicles(2024)

引用 0|浏览4
暂无评分
摘要
In this paper, we investigate the problem of optimizing the joint offloading policy in a distributed manner for multiple intelligent vehicles (IVs). During the journey in vehicular edge computing (VEC) networks, IVs continually optimize their joint offloading policy to minimize the longterm accumulated costs generated by executing computational tasks. The stochastic and repetitive interactions among IVs is modeled as a Markov game process. In this way, the optimization of the joint offloading policy is transformed to approximate a Markov perfect equilibrium in a general-sum Markov game. Moreover, we argue that training in the practical VEC networks using the classical centralized training and decentralized executing (CTDE) framework involves challenges of privacy and computational complexity. Motivated by these, we propose a reward-shared distributed policy optimization (RSDPO) method for the considered VEC networks to optimize the joint offloading policy. The experimental results demonstrate that the set of joint offloading policies using RSDPO approximates a Markov perfect equilibrium, and our RSDPO presents significant advantages in terms of converged latency and energy consumption compared with other methods.
更多
查看译文
关键词
Vehicular edge computing,joint offloading policy,distributed policy optimization,Markov game
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要