Gradient Coding with Clustering and Multi-Message Communication

2019 IEEE Data Science Workshop (DSW)(2019)

引用 23|浏览41
暂无评分
摘要
Gradient descent (GD) methods are commonly employed in machine learning problems to optimize the parameters of the model in an iterative fashion. For problems with massive datasets, computations are distributed to many parallel computing servers (i.e., workers) to speed up GD iterations. While distributed computing can increase the computation speed significantly, the per-iteration completion time is limited by the slowest straggling workers. Coded distributed computing can mitigate straggling workers by introducing redundant computations; however, existing coded computing schemes are mainly designed against persistent stragglers, and partial computations at straggling workers are discarded, leading to wasted computational capacity. In this paper, we propose a novel gradient coding (GC) scheme which allows multiple coded computations to be conveyed from each worker to the master per iteration. We numerically show that the proposed GC with multi-message communication (MMC) together with clustering provides significant improvements in the average completion time (of each iteration), with minimal or no increase in the communication load.
更多
查看译文
关键词
multimessage communication,gradient descent methods,machine learning problems,iterative fashion,parallel computing servers,GD iterations,computation speed,coded distributed computing,redundant computations,coded computing schemes,partial computations,computational capacity,communication load,gradient coding scheme
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要