Accelerating simulation of agent-based models on heterogeneous architectures.

GPGPU-6: Proceedings of the 6th Workshop on General Purpose Processor Using Graphics Processing Units(2013)

引用 19|浏览0
暂无评分
摘要
The wide usage of GPGPU programming models and compiler techniques enables the optimization of data-parallel programs on commodity GPUs. However, mapping GPGPU applications running on discrete parts to emerging integrated heterogeneous architectures such as the AMD Fusion APU and Intel Sandy/Ivy bridge with the CPU and the GPU on the same die has not been well studied. Classic time-step simulation applications represented by agent-based models have the intrinsic parallel structure that is a good fit for GPGPU architectures. However, when mapping these applications directly to the integrated GPUs, the performance may degrade due to less computation units and lower clock speed. This paper proposes an optimization to the GPGPU implementation of the agent-based model and illustrates it in the traffic simulation example. The optimization adapts the algorithm by moving part of the workload to the CPU to leverage the integrated architecture and the on-chip memory bus which is faster than the PCIe bus that connects the discrete GPU and the host. The experiments on discrete AMD Radeon GPU and AMD Fusion APU demonstrate that the optimization can achieve 1.08--2.71x performance speedup on the integrated architecture over the discrete platform.
更多
查看译文
关键词
heterogeneous architecture,accelerating simulation,discrete amd radeon gpu,gpgpu architecture,gpgpu implementation,gpgpu application,agent-based model,discrete gpu,discrete part,amd fusion apu,gpgpu programming model,integrated architecture,apu,gpgpu
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要