Accelerating Lossy Compression on HPC Datasets via Partitioning Computation for Parallel Processing.

HPCC/SmartCity/DSS(2019)

引用 3|浏览10
暂无评分
摘要
Recently, increasing attention has been paid to data reduction in the high-performance computing (HPC) environment where a large volume of data are produced continually during scientific simulations. A approach called SZ lossy compressor has been one of the best choices for HPC data reduction due to its high compression ratios while meeting data precision requirements. Currently, high compression rate is also strongly demanded because of fairly high data production throughput of many applications. In this work, we aim to accelerate the SZ compressor significantly by developing a parallel model in terms of the wildly-used point-wise relative error bound. It is non-trivial to parallelize SZ because of the strong data dependency in SZ. To address this issue, we develop a pipeline-like method and exploit a series of strategies to parallelize the 'logarithmic transformation' and 'prediction + quantization' stages for SZ. Our evaluation with real-world scientific simulation datasets shows that our design can accelerate the compression rate by over 2.0× in most cases while still guaranteeing the same compression ratio as the original serial version of SZ.
更多
查看译文
关键词
Lossy Compression,High-Performance Computing,Scientific Data,Compression Rate
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要