Addressing the Memory Bottleneck in AI Model Training

Ojika David, Patel Bhavesh, Reina G. Anthony, Boyer Trent, Martin Chad, Shah Prashant

arxiv(2020)

引用 0|浏览115
暂无评分
摘要
Using medical imaging as case-study, we demonstrate how Intel-optimized TensorFlow on an x86-based server equipped with 2nd Generation Intel Xeon Scalable Processors with large system memory allows for the training of memory-intensive AI/deep-learning models in a scale-up server configuration. We believe our work represents the first training of a deep neural network having large memory footprint (~ 1 TB) on a single-node server. We recommend this configuration to scientists and researchers who wish to develop large, state-of-the-art AI models but are currently limited by memory.
更多
查看译文
关键词
ai model training,memory bottleneck
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要