On the Whittle Index for Restless Multi-armed Hidden Markov Bandits.

IEEE Transactions on Automatic Control(2018)

引用 55|浏览20
暂无评分
摘要
We consider a restless multiarmed bandit in which each arm can be in one of two states. When an arm is sampled, the state of the arm is not available to the sampler. Instead, a binary signal with a known randomness that depends on the state of the arm is available. No signal is available if the arm is not sampled. An arm-dependent reward is accrued from each sampling. In each time step, each arm c...
更多
查看译文
关键词
Hidden Markov models,Markov processes,Computational modeling,Throughput,Numerical models
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要