Hybrid MPI/OpenMP Parallel Programming on Clusters of Multi-Core SMP Nodes

Weimar(2009)

引用 416|浏览0
暂无评分
摘要
Today most systems in high-performance computing (HPC) feature a hierarchical hardware design: Shared memory nodes with several multi-core CPUs are connected via a network infrastructure. Parallel programming must combine distributed memory parallelization on the node interconnect with shared memory parallelization inside each node. We describe potentials and challenges of the dominant programming models on hierarchically structured hardware: Pure MPI (message passing interface), pure OpenMP (with distributed shared memory extensions) and hybrid MPI+OpenMP in several flavors. We pinpoint cases where a hybrid programming model can indeed be the superior solution because of reduced communication needs and memory consumption, or improved load balance. Furthermore we show that machine topology has a significant impact on performance for all parallelization strategies and that topology awareness should be built into all applications in the future. Finally we give an outlook on possible standardization goals and extensions that could make hybrid programming easier to do with performance in mind.
更多
查看译文
关键词
memory parallelization,shared memory parallelization,load balance improvement,application program interfaces,openmp parallel programming,smp,parallel programming,memory consumption,high-performance computing,distributed memory systems,shared memory node,hybrid programming,hybrid mpi,multi-core smp nodes,resource allocation,message passing interface,hybrid programming model,shared memory systems,mpi-openmp parallel programming,multicore smp nodes,mpi,multi-core,message passing,openmp,dominant programming model,memory extension,distributed memory parallelization,data mining,programming model,programming,shared memory,topology,multi core,load balance,distributed shared memory,hardware,high performance computing,computational modeling,distributed memory
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要