Cooperative Deep Reinforcement Learning for Dynamic Pollution Plume Monitoring Using a Drone Fleet.

Mohamed Sami Assenine,Walid Bechkit,Ichrak Mokhtari,Hervé Rivano, Karima Benatchba

IEEE Internet Things J.(2024)

引用 0|浏览0
暂无评分
摘要
Monitoring pollution plumes is a key issue, given the harmful effects they cause. The dynamic of these plumes, which may be important due to meteorological conditions, makes their study difficult. Real-time monitoring in order to obtain an accurate mapping of the pollution dispersion is helpful and valuable to mitigate risks. In this work, we consider a fleet of cooperative drones carrying pollution sensors and operating in order to assess a pollution plume. The latter is assumed to follow a Gaussian Process (GP) with varying parameters. For this use case, we propose an efficient approach to characterize spatially and temporarily the plume while optimizing the path planning of drones. In our approach, drones are guided by a Deep Reinforcement Learning (DRL) model called Categorical Deep Q-Network (Categorical DQN) to maximize the plume coverage while considering budget constraints. Specifically, we develop a scalable Independent Q-Learning (IQL) scheme that shares team rewards based on each drone’s deployment relevance and therefore ensures cooperation. We evaluate the performance of the plume parameter estimation as well as the maps generated by the GP regression. By testing our framework on several plume scenarios, we show that it offers good results in terms of both estimation quality and run-time efficiency.
更多
查看译文
关键词
Drone fleet,Pollution plume monitoring,Reinforcement Learning,Independent Q-Learning,Gaussian Process,Multi-agent Informative Path Planning
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要