Density-Based Data Pruning Method for Deep Reinforcement Learning

Teerapat Rojanaarpa,Irina Kataeva

2016 15th IEEE International Conference on Machine Learning and Applications (ICMLA)(2016)

引用 4|浏览0
暂无评分
摘要
We present a density-based Data Pruning method for Deep Reinforcement Learning (DRL) to improve learning stability and long-term memory in rare situations. The method controls density distribution in the experience pool by discarding high correlation data and preserving rare and unique data. We apply our method to Deep Q-networks (DQN) and Deep Deterministic Policy Gradients (DDPG) for testing in discrete and continuous action space, respectively. We evaluate our method on path following tasks in a simulated physical environment. Compared to other conventional methods such as First-In-First-Out (FIFO), our method provides a significant improvement in performance and learning stability, the average cumulative reward is increased by up to 21% and the standard deviation of the cumulative reward over multiple trials is reduced by 80%. In addition, long-term memory improvement is shown as the agent can remember and perform a behavior corresponding to a past rare event.
更多
查看译文
关键词
density-based data pruning method,deep reinforcement learning,learning stability,long-term memory,density distribution,high correlation data,deep q-networks,deep deterministic policy gradients,discrete action space,continuous action space,simulated physical environment,long-term memory improvement
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要