DocumentCode
3322345
Title
High Throughput Intra-Node MPI Communication with Open-MX
Author
Goglin, Brice
Author_Institution
INRIA Bordeaux, LaBRI, Talence
fYear
2009
fDate
18-20 Feb. 2009
Firstpage
173
Lastpage
180
Abstract
The increasing number of cores per node in high-performance computing requires an efficient intra-node MPI communication subsystem. Most existing MPI implementations rely on two copies across a shared memory-mapped file. Open-MX offers a single-copy mechanism that is tightly integrated in its regular communication stack, making it transparently available to the MX backend of many MPI layers. We describe this implementation and its offloaded copy backend using I/OAT hardware. Memory pinning requirements are then discussed, and overlapped pinning is introduced to enable the start of Open-MX intra-node data transfer earlier. Performance evaluation shows that this local communication stack performs better than MPICH2 and Open-MPI for large messages, reaching up to 70% better throughput in micro-benchmarks when using I/OAT copy offload. Thanks to a single-copy being involved, the Open-MX intra-node communication throughput also does not heavily depend on cache sharing between processing cores, making these performance improvements easier to observe in real applications.
Keywords
application program interfaces; message passing; storage allocation; Open-MX intra-node data transfer; high throughput intra-node MPI communication subsystem; high-performance computing; memory pinning; overlapped pinning; single-copy mechanism; Collaborative software; Computer architecture; Computer networks; Concurrent computing; Delay; Distributed computing; Hardware; Multicore processing; Performance evaluation; Throughput;
fLanguage
English
Publisher
ieee
Conference_Titel
Parallel, Distributed and Network-based Processing, 2009 17th Euromicro International Conference on
Conference_Location
Weimar
ISSN
1066-6192
Print_ISBN
978-0-7695-3544-9
Type
conf
DOI
10.1109/PDP.2009.20
Filename
4912931
Link To Document