DocumentCode :
641330
Title :
Aggrandizing Hadoop in terms of node Heterogeneity & Data Locality
Author :
Sujitha, S. ; Jaganathan, Suresh
Author_Institution :
Dept. of Comput. Sci. & Eng., Sri Sivasubramania Nadar Coll. of Eng., Chennai, India
fYear :
2013
fDate :
28-29 March 2013
Firstpage :
145
Lastpage :
151
Abstract :
The growth of data has increased exponentially in recent years. In this context, data-center scale computer systems are built to meet the high storage and processing demands of these applications. Such systems are composed of hundreds, thousands, or even millions of commodity computers connected through a LAN housed in a data center. It has a much larger scale than a traditional computer cluster. Hadoop enables the distributed processing of large data sets across clusters of commodity servers. It is designed to scale up from a single server to thousands of machines, with a very high degree of fault tolerance. Strength of Hadoop is in its ability to detect and handle failures. The original Hadoop native task scheduler implicitly assumes that cluster nodes are homogeneous. This assumption is used to identify a slow task and re-execute it. However, this assumption does not hold where the cluster nodes are heterogeneous, since speculatively identifying a slow task will give rise to erroneous conclusions. In MapReduce, the sub-task is transferred to a node for execution. The input to the subtask, if not present in the node, must be transferred from another node in the network. Transferring data takes time and delays execution. In this paper, we have proposed a methodology for improving Hadoop in-terms of Heterogeneity and Data Locality. The performance of improved version can be measured using these metrics, i)execution time, ii)response time, iii)tasks submitted, iv)time related to jobs i.e. arrival, start and completion, v)completed task, vi)fairness, vii)locality and viii)mean completion time.
Keywords :
computer centres; distributed processing; fault tolerant computing; Hadoop; MapReduce; commodity servers; completed task; data center scale computer systems; data locality; distributed processing; execution time; fairness; fault tolerance; large data sets; mean completion time; native task scheduler; node heterogeneity; response time; tasks submitted; time related to jobs; Computers; Monitoring; Big Data; Cloud Computing; Data Computing; Heterogeneity; Schedulers;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Smart Structures and Systems (ICSSS), 2013 IEEE International Conference on
Conference_Location :
Chennai
Print_ISBN :
978-1-4673-6240-5
Type :
conf
DOI :
10.1109/ICSSS.2013.6623017
Filename :
6623017
Link To Document :
بازگشت