Title :
In search of the best MPI-OpenMP distribution for optimum Intel-MIC cluster performance
Author :
Utrera, Gladys ; Gil, Marisa ; Martorell, Xavier
Author_Institution :
Comput. Archit. Dept., Univ. Politec. de Catalunya-BarcelonaTECH, Barcelona, Spain
Abstract :
Applications for HPC platforms are mainly based on hybrid programming models: MPI for communication and OpenMP for task and fork-join parallelism to exploit shared memory communication inside a node. On the basis of this scheme, much research has been carried out to improve performance. Some examples are: the overlap of communication and computation, or the increase of speedup and bandwidth on new network fabrics (i.e. Infiniband and 10GB or 40GB ethernet). Henceforth, as far as computation and communication are concerned, the HPC platforms will be heterogeneous with high-speed networks. And, in this context, an important issue is to decide how to distribute the workload among all the nodes in order to balance the application execution as well as choosing the most appropriate programming model to exploit parallelism inside the node. In this paper we propose a mechanism to balance dynamically the work distribution among the heterogeneous components of an heterogeneous cluster based on their performance characteristics. For our evaluations we run the miniFE mini-application of the Mantevo suite benchmark, in a heterogeneous Intel MIC cluster. Experimental results show that making an effort to choose the appropriate number of threads can improve performance significantly over choosing the maximum available number of cores in the Intel MIC.
Keywords :
application program interfaces; message passing; parallel processing; Ethernet; Infiniband; Intel MIC; MPI-OpenMP distribution; Mantevo suite benchmark; dynamically balanced work distribution; fork-join parallelism; heterogeneous HPC platforms; heterogeneous Intel MIC cluster; high-speed networks; hybrid programming models; miniFE miniapplication; network fabrics; optimum Intel-MIC cluster performance; performance improvement; programming model; shared memory communication; task parallelism; workload distribute; Benchmark testing; Computational modeling; Libraries; Microwave integrated circuits; Parallel processing; Programming; Scalability; Heterogeneous HPC platforms; Hybrid MPI-OpenMP; Intel MIC; MPI_Allreduce;
Conference_Titel :
High Performance Computing & Simulation (HPCS), 2015 International Conference on
Conference_Location :
Amsterdam
Print_ISBN :
978-1-4673-7812-3
DOI :
10.1109/HPCSim.2015.7237072