Decomposition-based Multi-Agent Distributional Reinforcement Learning for Task-Oriented UAV Collaboration with Noisy Rewards.

2023 International Conference on Wireless Communications and Signal Processing (WCSP)(2023)

Cited 0|Views9
No score
Abstract
Collaborated unmanned aerial vehicles (UAVs) are often deployed to perform complex tasks on the basis of multi-agent reinforcement learning (MARL). However, the environmental disturbance, commonly leading to noisy observations (e.g., rewards, states), could significantly shape the performance of task-oriented UAV collaboration. Therefore, it becomes imperative to revolutionize the design of MARL, so as to capably ameliorate the annoying impact of noisy rewards. In this paper, we propose a novel decomposition-based multi-agent distributional RL method by approximating the globally shared noisy reward by a Gaussian mixture model and decomposing it into the combination of individual distributional local rewards, with which each agent can be updated locally through distributional RL. Besides, the optimality of the distributional decomposition is theoretically validated, while the design of loss functions is carefully calibrated to tackle the decomposition ambiguity. We also verify the effectiveness of the proposed method through extensive simulation experiments with noisy rewards.
More
Translated text
Key words
Multi-agent reinforcement learning,noisy environment,distributional reinforcement learning,distribution decomposition
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined