An Empirical and Conceptual Categorization of Value-based Exploration Methods

semanticscholar(2019)

引用 0|浏览1
暂无评分
摘要
This paper categorizes several exploration methods for reinforcement learning according to their underlying heuristics. Using linear Q-learning, we compare representative methods from each category of methods on a set of domains designed to pose different exploratory challenges. We find that the relative performance of each method depends on the specific exploratory challenge posed by the domain. Our results suggest that each exploration heuristic encodes a bias which is appropriate for a subset of environments.
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要