DocumentCode :
1999638
Title :
High-Performance RDMA-based Design of Hadoop MapReduce over InfiniBand
Author :
Wasi-ur-Rahman, Md ; Islam, Nusrat Sharmin ; Xiaoyi Lu ; Jose, Jithin ; Subramoni, Hari ; Hao Wang ; Panda, Dhabaleswar K. Dk
Author_Institution :
Dept. of Comput. Sci. & Eng., Ohio State Univ., Columbus, OH, USA
fYear :
2013
fDate :
20-24 May 2013
Firstpage :
1908
Lastpage :
1917
Abstract :
MapReduce is a very popular programming model used to handle large datasets in enterprise data centers and clouds. Although various implementations of MapReduce exist, Hadoop MapReduce is the most widely used in large data centers like Facebook, Yahoo! and Amazon due to its portability and fault tolerance. Network performance plays a key role in determining the performance of data intensive applications using Hadoop MapReduce as data required by the map and reduce processes can be distributed across the cluster. In this context, data center designers have been looking at high performance interconnects such as InfiniBand to enhance the performance of their Hadoop MapReduce based applications. However, achieving better performance through usage of high performance interconnects like InfiniBand is a significant task. It requires a careful redesign of communication framework inside MapReduce. Several assumptions made for current socket based communication in the current framework do not hold true for high performance interconnects. In this paper, we propose the design of an RDMA-based Hadoop MapReduce over InfiniBand and several design elements: data shuffle over InfiniBand, in-memory merge mechanism for the Reducer, and pre-fetch data for the Mapper. We perform our experiments on native InfiniBand using Remote Direct Memory Access (RDMA) and compare our results with that of Hadoop-A [1] and default Hadoop over different interconnects and protocols. For all these experiments, we perform network level parameter tuning and use optimum values for each Hadoop design. Our performance results show that, for a 100GB TeraSort running on an eight node cluster, we achieve a performance improvement of 32% over IP-over InfiniBand (IPoIB) and 21% over Hadoop-A. With multiple disks per node, this benefit rises up to 39% over IPoIB and 31% over Hadoop-A.
Keywords :
computer network performance evaluation; distributed databases; fault tolerant computing; file organisation; merging; public domain software; Hadoop MapReduce; Hadoop-A; IP-over InfiniBand; IPoIB; TeraSort; big data analytic applications; communication framework; data intensive applications; data shuffle; default Hadoop; enterprise data centers; fault tolerance; high-performance RDMA-based design; in-memory merge mechanism; mapper; network level parameter tuning; network performance; open-source frameworks; performance interconnects; portability; pre-fetch data; programming model; protocols; reducer; remote direct memory access; Benchmark testing; Corporate acquisitions; Distributed databases; IP networks; Information management; Libraries; Software;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Parallel and Distributed Processing Symposium Workshops & PhD Forum (IPDPSW), 2013 IEEE 27th International
Conference_Location :
Cambridge, MA
Print_ISBN :
978-0-7695-4979-8
Type :
conf
DOI :
10.1109/IPDPSW.2013.238
Filename :
6651094
Link To Document :
بازگشت