DocumentCode :
120720
Title :
Hadoop: Addressing challenges of Big Data
Author :
Singh, Karam ; Kaur, Rupinderjit
Author_Institution :
CSE Dept., Lovely Prof. Univ., Jalandhar, India
fYear :
2014
fDate :
21-22 Feb. 2014
Firstpage :
686
Lastpage :
689
Abstract :
Hadoop is an open source cloud computing platform of the Apache Foundation that provides a software programming framework called MapReduce and distributed file system, HDFS. It is a Linux based set of tools that uses commodity hardware, which are relatively inexpensive, to handle, analyze and transform large quantity of data. Hadoop Distributed File System, HDFS, stores huge data set reliably and streams it to user application at high bandwidth and MapReduce is a framework that is used for processing massive data sets in a distributed fashion over a several machines. This paper gives a brief overview of Big Data, Hadoop MapReduce and Hadoop Distributed File System along with its architecture.
Keywords :
Big Data; Linux; cloud computing; distributed databases; public domain software; Apache Foundation; Big Data; HDFS; Hadoop distributed file system; Linux; MapReduce; commodity hardware; open source cloud computing platform; software programming framework; Data handling; Data storage systems; Distributed databases; File systems; Hardware; Information management; Reliability; Big Data; Datanode; HDFS; Hadoop; MapReduce; Namenode;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Advance Computing Conference (IACC), 2014 IEEE International
Conference_Location :
Gurgaon
Print_ISBN :
978-1-4799-2571-1
Type :
conf
DOI :
10.1109/IAdCC.2014.6779407
Filename :
6779407
Link To Document :
بازگشت