Understanding Training Efficiency of Deep Learning Recommendation Models at Scale

2021 IEEE International Symposium on High-Performance Computer Architecture (HPCA)(2021)

引用 81|浏览158
暂无评分
摘要
The use of GPUs has proliferated for machine learning workflows and is now considered mainstream for many deep learning models. Meanwhile, when training state-of-the-art personal recommendation models, which consume the highest number of compute cycles at our large-scale datacenters, the use of GPUs came with various challenges due to having both compute-intensive and memory-intensive components. ...
更多
查看译文
关键词
Training,Deep learning,Computational modeling,Memory management,Graphics processing units,Production,Throughput
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要