Avoiding Synchronization In First-Order Methods For Sparse Convex Optimization

international parallel and distributed processing symposium(2018)

引用 8|浏览53
暂无评分
摘要
Parallel computing has played an important role in speeding up convex optimization methods for big data analytics and large-scale machine learning (ML). However, the scalability of these optimization methods is inhibited by the cost of communicating and synchronizing processors in a parallel setting. Iterative ML methods are particularly sensitive to communication cost since they often require communication every iteration. In this work, we extend well-known techniques from Communication-Avoiding Krylov subspace methods to first-order, block coordinate descent methods for Support Vector Machines and Proximal Least-Squares problems. Our Synchronization-Avoiding (SA) variants reduce the latency cost by a tunable factor of s at the expense of a factor of s increase in flops and bandwidth costs. We show that the SA-variants are numerically stable and can attain large speedups of up to 5:1x on a Cray XC30 supercomputer.
更多
查看译文
关键词
Synchronization-Avoiding, Support Vector Machines, Proximal Least-Squares, Sparse Convex Optimization, Coordinate Descent Methods
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要