Author:
Martín-Álvarez Iker,Aliaga José I.,Castillo Maribel,Iserte Sergio
Abstract
AbstractApplying malleability to HPC systems can increase their productivity without degrading or even improving the performance of running applications. This paper presents Proteo, a configurable framework that allows to design benchmarks to study the effect of malleability on a system, and also incorporates malleability into a real application. Proteo consists of two modules: SAM allows to emulate the computational behavior of iterative scientific MPI applications, and MaM is able to reconfigure a job during execution, adjusting the number of processes, redistributing data, and resuming execution. An in-depth study of all the possibilities shows that Proteo is able to behave like a real malleable or non-malleable application in the range [0.85, 1.15]. Furthermore, the different methods defined in MaM for process management and data redistribution are analyzed, concluding that asynchronous malleability, where reconfiguration and application execution overlap, results in a 1.15$$\times$$
×
speedup.
Funder
Ministerio de Ciencia e Innovación
Generalitat Valenciana
European Commission
Universitat Jaume I
Publisher
Springer Science and Business Media LLC
Reference58 articles.
1. Dongarra J, Luszczek P (2011) In: Padua D (ed) TOP500. Springer, Boston, pp 2055–2057. https://doi.org/10.1007/978-0-387-09766-4_157
2. Message Passing Interface Forum: MPI: a message-passing interface standard version 4.1. (2023). https://www.mpi-forum.org/docs/mpi-4.1/mpi41-report.pdf
3. Hori A, Yoshinaga K, Herault T, Bouteiller A, Bosilca G, Ishikawa Y (2020) Overhead of using spare nodes. Int J High Perform Comput Appl 34(2):208–226. https://doi.org/10.1177/1094342020901885
4. Feitelson DG (1996) Packing schemes for gang scheduling. Lecture notes in computer science book series (LNCS), vol 1162. Springer, Heidelberg, pp 89–110
5. Bernholdt DE, Boehm S, Bosilca G, Gorentla Venkata M, Grant RE, Naughton T, Pritchard HP, Schulz M, Vallee GR (2020) A survey of MPI usage in the US exascale computing project. Concurr Comput Pract Exp 32(3):4851. https://doi.org/10.1002/cpe.4851