Deep Learning at Scale

2019 27th Euromicro International Conference on Parallel, Distributed and Network-Based Processing (PDP)(2019)

引用 6|浏览87
暂无评分
摘要
This work presents a novel approach to distributed training of deep neural networks (DNNs) that aims to overcome the issues related to mainstream approaches to data parallel training. Established techniques for data parallel training are discussed from both a parallel computing and deep learning perspective, then a different approach is presented that is meant to allow DNN training to scale while retaining good convergence properties. Moreover, an experimental implementation is presented as well as some preliminary results.
更多
查看译文
关键词
Training,Parallel processing,Neural networks,Convergence,Computational modeling,Data models,Deep learning
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要