Affiliation:
1. Texas A&M University, College Station, TX
Abstract
The NAS Parallel Benchmarks (NPB) are well-known applications with the fixed algorithms for evaluating parallel systems and tools. Multicore supercomputers provide a natural programming paradigm for hybrid programs, whereby OpenMP can be used with the data sharing with the multicores that comprise a node and MPI can be used with the communication between nodes. In this paper, we use SP and BT benchmarks of MPI NPB 3.3 as a basis for a comparative approach to implement hybrid MPI/OpenMP versions of SP and BT. In particular, we can compare the performance of the hybrid SP and BT with the MPI counterparts on large-scale multicore supercomputers. Our performance results indicate that the hybrid SP outperforms the MPI SP by up to 20.76%, and the hybrid BT outperforms the MPI BT by up to 8.58% on up to 10,000 cores on BlueGene/P at Argonne National Laboratory and Jaguar (Cray XT4/5) at Oak Ridge National Laboratory. We also use performance tools and MPI trace libraries available on these supercomputers to further investigate the performance characteristics of the hybrid SP and BT.
Publisher
Association for Computing Machinery (ACM)
Subject
Computer Networks and Communications,Hardware and Architecture,Software
Reference11 articles.
1. Argonne Leadership Computing Facility BlueGene/P (Intrepid) Argonne National Laboratory http://www.alcf.anl.gov/resources. Argonne Leadership Computing Facility BlueGene/P (Intrepid) Argonne National Laboratory http://www.alcf.anl.gov/resources.
Cited by
16 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Supercharging the APGAS Programming Model with Relocatable Distributed Collections;Scientific Programming;2022-09-21
2. First Experiences in Performance Benchmarking with the New SPEChpc 2021 Suites;2022 22nd IEEE International Symposium on Cluster, Cloud and Internet Computing (CCGrid);2022-05
3. IMPACC;Proceedings of the 25th ACM International Symposium on High-Performance Parallel and Distributed Computing;2016-05-31
4. MPI+ULT: Overlapping Communication and Computation with User-Level Threads;2015 IEEE 17th International Conference on High Performance Computing and Communications, 2015 IEEE 7th International Symposium on Cyberspace Safety and Security, and 2015 IEEE 12th International Conference on Embedded Software and Systems;2015-08
5. Characterizing MPI and Hybrid MPI+Threads Applications at Scale: Case Study with BFS;2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing;2015-05