An Exact Distributed Newton Method For Reinforcement Learning

2016 IEEE 55TH CONFERENCE ON DECISION AND CONTROL (CDC)(2016)

引用 5|浏览47
暂无评分
摘要
In this paper, we propose a distributed second order method for reinforcement learning. Our approach is the fastest in literature so-far as it outperforms state-of-the-art methods, including ADMM, by significant margins. We achieve this by exploiting the sparsity pattern of the dual Hessian and transforming the problem of computing the Newton direction to one of solving a sequence of symmetric diagonally dominant system of equations. We validate the above claim both theoretically and empirically. On the theoretical side, we prove that similar to exact Newton, our algorithm exhibits super-linear convergence within a neighborhood of the optimal solution. Empirically, we demonstrate the superiority of this new method on a set of benchmark reinforcement learning tasks.
更多
查看译文
关键词
Learning (artificial intelligence),Trajectory,Convergence,Approximation algorithms,Optimization,Newton method,Robots
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要