Distributed Optimization for Quadratic Cost Functions over Large-Scale Networks with Quantized Communication and Finite-Time Convergence

CoRR(2023)

引用 0|浏览7
暂无评分
摘要
We propose two distributed iterative algorithms that can be used to solve, in finite time, the distributed optimization problem over quadratic local cost functions in large-scale networks. The first algorithm exhibits synchronous operation whereas the second one exhibits asynchronous operation. Both algorithms share salient features. Specifically, the algorithms operate exclusively with quantized values, which means that the information stored, processed and exchanged between neighboring nodes is subject to deterministic uniform quantization. The algorithms rely on event-driven updates in order to reduce energy consumption, communication bandwidth, network congestion, and/or processor usage. Finally, once the algorithms converge, nodes distributively terminate their operation. We prove that our algorithms converge in a finite number of iterations to the exact optimal solution depending on the quantization level, and we present applications of our algorithms to (i) optimal task scheduling for data centers, and (ii) global model aggregation for distributed federated learning. We provide simulations of these applications to illustrate the operation, performance, and advantages of the proposed algorithms. Additionally, it is shown that our proposed algorithms compare favorably to algorithms in the current literature. Quantized communication and asynchronous updates increase the required time to completion, but finite-time operation is maintained.
更多
查看译文
关键词
quadratic cost functions,quantized communication,optimization,networks,large-scale,finite-time
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要