MRPB: Memory request prioritization for massively parallel processors

HPCA(2014)

引用 205|浏览91
暂无评分
摘要
Massively parallel, throughput-oriented systems such as graphics processing units (GPUs) offer high performance for a broad range of programs. They are, however, complex to program, especially because of their intricate memory hierarchies with multiple address spaces. In response, modern GPUs have widely adopted caches, hoping to providing smoother reductions in memory access traffic and latency. Unfortunately, GPU caches often have mixed or unpredictable performance impact due to cache contention that results from the high thread counts in GPUs. We propose the memory request prioritization buffer (MRPB) to ease GPU programming and improve GPU performance. This hardware structure improves caching efficiency of massively parallel workloads by applying two prioritization methods-request reordering and cache bypassing-to memory requests before they access a cache. MRPB then releases requests into the cache in a more cache-friendly order. The result is drastically reduced cache contention and improved use of the limited per-thread cache capacity. For a simulated 16KB L1 cache, MRPB improves the average performance of the entire PolyBench and Rodinia suites by 2.65× and 1.27× respectively, outperforming a state-of-the-art GPU cache management technique.
更多
查看译文
关键词
polybench suites,hardware structure,address spaces,parallel processing,gpu caches,mrpb,cache storage,request reordering,massively parallel processors,memory request prioritization buffer,graphics processing units,gpu programming,memory request prioritization,massively parallel throughput-oriented systems,memory hierarchies,cache bypassing,memory access latency,performance evaluation,gpu performance,simulated l1 cache,thread counts,caching efficiency,rodinia suites,prioritization methods,limited per-thread cache capacity,memory access traffic,gpu cache management technique,instruction sets,throughput,kernel,pipelines,hardware
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要