First Order Methods with Markovian Noise: from Acceleration to Variational Inequalities
arxiv(2023)
摘要
This paper delves into stochastic optimization problems that involve
Markovian noise. We present a unified approach for the theoretical analysis of
first-order gradient methods for stochastic optimization and variational
inequalities. Our approach covers scenarios for both non-convex and strongly
convex minimization problems. To achieve an optimal (linear) dependence on the
mixing time of the underlying noise sequence, we use the randomized batching
scheme, which is based on the multilevel Monte Carlo method. Moreover, our
technique allows us to eliminate the limiting assumptions of previous research
on Markov noise, such as the need for a bounded domain and uniformly bounded
stochastic gradients. Our extension to variational inequalities under Markovian
noise is original. Additionally, we provide lower bounds that match the oracle
complexity of our method in the case of strongly convex optimization problems.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要