Bridging the Reality Gap Between Virtual and Physical Environments Through Reinforcement Learning.

IEEE Access(2023)

引用 0|浏览11
暂无评分
摘要
Creating Reinforcement learning(RL) agents that can perform tasks in the real-world robotic systems remains a challenging task due to inconsistencies between the virtual-and the real-world. This is known as the "reality-gap " which hinders the performance of a RL agent trained in a virtual environment. The research describes the techniques used to train the models, generate randomized environments, reward function, and techniques utilized to transfer the model to the physical environment for evaluation. For this investigation, a low-cost 3-degrees-of-freedom (DOF) Steward platform was 3D modeled and created virtually and physically. The goal of the 3D-Stewart platform was to guide and balance the marble towards the center. Custom end-to-end APIs were developed to interact with the Godot game engine, manipulate physics and dynamics, interact with the in-game lighting and perform environment randomizations. Two RL algorithms: Q-learning and Actor-Critic, were implemented to evaluate the performance by using domain randomization and induced noise to bridge the reality gap. For Q-learning, raw frames were used to make predictions while Actor-Critic utilized marble position, velocity vector and relative position by pre-processing captured frames. The experimental results show the effectiveness of domain randomization and introduction of noise during the training.
更多
查看译文
关键词
Robots,Adaptation models,Virtual environments,Training,Q-learning,Transfer learning,Reinforcement learning,Actor-critic,deep Q-learning,reinforcement learning,transfer-learning,sim-to-real,robotics
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要