Coverage Path Planning with Proximal Policy Optimization in a Grid-based Environment.

Aleksandr Ianenko, Alexander Artamonov, Georgii Sarapulov, Alexey Safaraleev, Sergey Bogomolov,Dong-Ki Noh

CDC(2020)

引用 3|浏览0
暂无评分
摘要
We present an on-line approach for coverage path planning in 2D grid environments based on reinforcement learning. We used actor-critic architecture with convolutional layers to learn an agent's policy from simulated limited-range sensor observations. We experimented with different reward functions and network architectures to get a minimal repetition rate. Our results show that model generalizes well to unseen environments with complex geometry and dynamic obstacles and demonstrates the ability to learn some optimal trajectory patterns like circular and boustrophedon motion. An approach may as well be suitable for a multi-agent setting with minor adjustments, as shown by our simulations.
更多
查看译文
关键词
coverage path planning,proximal policy optimization,grid-based environment,on-line approach,2D grid environments,reinforcement learning,actor-critic architecture,convolutional layers,agent policy,reward functions,network architectures,minimal repetition rate,unseen environments,optimal trajectory patterns,limited-range sensor observations,complex geometry,dynamic obstacles,circular motion,boustrophedon motion,multiagent setting
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要