Convergence Rate Analysis for Distributed Optimization with Localization

2019 57th Annual Allerton Conference on Communication, Control, and Computing (Allerton)(2019)

引用 0|浏览26
暂无评分
摘要
We study the effect of the localization scheme on the convergence rate in the setting of distributed gradient descent for smooth optimization. Localization is a technique that exploits the partial dependency structure in the objective functions to reduce the memory storage and communication involved in distributed optimization algorithms. We find that the localization scheme could reduce the convergence rate by a constant factor, which depends on the sizes of the local networks and the spectral radii of the associated optimal doubly stochastic matrices for averaging. The choice of local networks directly affects the memory cost, the communication cost, and the convergence speed. In order to characterize the relation, we explore how the optimal spectral radius depends on the topology of a graph, i.e., local network in the context. We provide a numerical example that shows the saving in terms of memory, communication, as well as iterations required could be tremendous with the localization schemes.
更多
查看译文
关键词
associated optimal doubly stochastic matrices,distributed optimization algorithms,memory storage,partial dependency structure,smooth optimization,distributed gradient descent,convergence rate analysis,localization scheme,local network,optimal spectral radius,convergence speed,communication cost,memory cost
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要