Ray RLlib: A Framework for Distributed Reinforcement Learning

arXiv: Artificial Intelligence(2017)

引用 24|浏览52
暂无评分
摘要
Reinforcement learning (RL) training involves the deep nesting of highly irregular computation patterns, each of which typically exhibits opportunities for distributed computation. Current RL libraries offer parallelism at the level of the entire program, coupling all algorithm components together and making existing implementations difficult to scale, combine, and reuse. argue for distributing RL components in a composable way by adapting algorithms for top-down hierarchical control, thereby encapsulating parallelism and resource requirements within short-running compute tasks. We demonstrate this principle by building RLlib on top of a task-based framework and show that we can implement a wide range of state-of-the art algorithms on top of a small set of general abstractions. These abstractions are key to composability and reuse in RLlib and do not come at the cost of performance---in our experiments, RLlib matches or exceeds the performance of highly optimized reference implementations. Ray RLlib is available as part of Ray at this https URL
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要