Bandits and Experts in Metric Spaces.

Journal of the ACM (JACM)(2019)

引用 77|浏览153
暂无评分
摘要
In a multi-armed bandit problem, an online algorithm chooses from a set of strategies in a sequence of trials to maximize the total payoff of the chosen strategies. While the performance of bandit ...
更多
查看译文
关键词
Multi-armed bandits,regret,online learning,metric spaces,covering dimension,Lipschitz-continuity
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要