Aggregating Local Storage for Scalable Deep Learning I/O

2019 IEEE/ACM Third Workshop on Deep Learning on Supercomputers (DLS)(2019)

引用 7|浏览24
暂无评分
摘要
Deep learning applications introduce heavy I/O loads on computer systems. The inherently long-running, highly concurrent, and random file accesses can easily saturate traditional shared file systems and negatively impact other users. We investigate here a solution to these problems based on leveraging local storage and the interconnect to serve training datasets at scale. We present FanStore, a user-level transient object store that provides low-latency and scalable POSIX-compliant file access by integrating the function interception technique and various metadata/data placement strategies. On a single node, FanStore provides performance similar to that of the XFS journaling file system. On many nodes, our experiments with real applications show that FanStore achieves over 90% scaling efficiency.
更多
查看译文
关键词
scalable deep,deep learning applications,computer systems,random file accesses,traditional shared file systems,interconnect,training datasets,local storage leveraging,scaling efficiency,XFS journaling file system,FanStore,function interception technique,POSIX-compliant file access,user-level transient object store
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要