An Empirical and Conceptual Categorization of Value-based Exploration Methods
semanticscholar(2019)
摘要
This paper categorizes several exploration methods for reinforcement learning according to their underlying heuristics. Using linear Q-learning, we compare representative methods from each category of methods on a set of domains designed to pose different exploratory challenges. We find that the relative performance of each method depends on the specific exploratory challenge posed by the domain. Our results suggest that each exploration heuristic encodes a bias which is appropriate for a subset of environments.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要