Modeling Interaction-Aware Driving Behavior using Graph-Based Representations and Multi-Agent Reinforcement Learning.

Fabian Konstantinidis,Moritz Sackmann,Ulrich Hofmann, Christoph Stiller

2023 IEEE 26th International Conference on Intelligent Transportation Systems (ITSC)(2023)

引用 0|浏览0
暂无评分
摘要
Modeling the driving behavior of traffic partici-pants in highly interactive traffic situations, such as roundabouts, poses a significant challenge due to the complex interactions and the variety of traffic situations. To address this task, we propose a combination of graph-based representations of the environment with Multi-Agent Reinforcement Learning (MARL). By utilizing a graph-based representation of the local environment of each vehicle, our approach efficiently accounts for road structures and a varying number of surrounding vehicles interacting with each other. Building upon this representation, MARL enables us to learn a driving policy based on a minimal set of principles: drivers want to move along the road while avoiding collisions and maintaining comfortable accelerations. Sharing the learned policy among all agents allows us to leverage Proximal Policy Optimization (PPO), a policy gradient Reinforcement Learning (RL) algorithm. To evaluate our proposed model, we conduct experiments in a roundabout scenario from the INTERACTION dataset and compare it to a model learned via Behavior Cloning (BC). The results demonstrate that our proposed model is capable of maneuvering through dense traffic, indicating that our graph-based representation is well suitable for modeling and understanding complex road layouts and interactions between agents.
更多
查看译文
关键词
Multi-agent Reinforcement Learning,Graph-based Representation,Collision,Local Environment,Policy Learning,Representation Of The Environment,Interaction Datasets,Interaction Situations,Traffic Situation,Proximal Policy Optimization,Real Situation,Simulation Environment,Learning Behavior,Reward Function,High Reward,Linear Layer,Environmental Agents,Static Environment,Reinforcement Learning Model,Shared Environment,Polyline,Partially Observable Markov Decision Process,Road Markings,Rule-based Model,Value Function Approximation,Environment In The Form,Expert Demonstrations,Driver Model,Lane Center,Traffic Scenarios
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要