Exponentially fast convergence to (strict) equilibrium via hedging.

arXiv: Computer Science and Game Theory(2016)

引用 23|浏览18
暂无评分
摘要
Motivated by applications to data networks where fast convergence is essential, we analyze the problem of learning in generic N-person games that admit a Nash equilibrium in pure strategies. Specifically, we consider a scenario where players interact repeatedly and try to learn from past experience by small adjustments based on local - and possibly imperfect - payoff information. For concreteness, we focus on the so-called hedge variant of the exponential weights algorithm where players select an action with probability proportional to the exponential of the actionu0027s cumulative payoff over time. When players have perfect information on their mixed payoffs, the algorithm converges locally to a strict equilibrium and the rate of convergence is exponentially fast - of the order of $mathcal{O}(exp(-asum_{j=1}^{t}gamma_{j}))$ where $au003e0$ is a constant and $gamma_{j}$ is the algorithmu0027s step-size. In the presence of uncertainty, convergence requires a more conservative step-size policy, but with high probability, the algorithm remains locally convergent and achieves an exponential convergence rate.
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要