Multi-gpu training and parallel cpu computing for the machine learning experiments using ariadne library

P. Goncharov, A. Nikolskaia,G. Ososkov, E. Rezvaya,D. Rusov,E. Shchavelev

9th International Conference "Distributed Computing and Grid Technologies in Science and Education"(2021)

引用 0|浏览9
暂无评分
摘要
Modern machine learning (ML) tasks and neural network (NN) architectures require huge amounts ofGPU computational facilities and demand high CPU parallelization for data preprocessing. At thesame time, the Ariadne library, which aims to solve complex high-energy physics tracking tasks withthe help of deep neural networks, lacks multi-GPU training and efficient parallel data preprocessing onthe CPU.In our work, we present our approach for the Multi-GPU training in the Ariadne library. We willpresent efficient data-caching, parallel CPU data preprocessing, generic ML experiment setup forprototyping, training, and inference deep neural network models. Results in terms of speed-up andperformance for the existing neural network approaches are presented with the help of GOVORUNcomputing resources.
更多
查看译文
关键词
ariadne library,parallel cpu computing,machine learning experiments,multi-gpu
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要