Title :
Hadoop: Addressing challenges of Big Data
Author :
Singh, Karam ; Kaur, Rupinderjit
Author_Institution :
CSE Dept., Lovely Prof. Univ., Jalandhar, India
Abstract :
Hadoop is an open source cloud computing platform of the Apache Foundation that provides a software programming framework called MapReduce and distributed file system, HDFS. It is a Linux based set of tools that uses commodity hardware, which are relatively inexpensive, to handle, analyze and transform large quantity of data. Hadoop Distributed File System, HDFS, stores huge data set reliably and streams it to user application at high bandwidth and MapReduce is a framework that is used for processing massive data sets in a distributed fashion over a several machines. This paper gives a brief overview of Big Data, Hadoop MapReduce and Hadoop Distributed File System along with its architecture.
Keywords :
Big Data; Linux; cloud computing; distributed databases; public domain software; Apache Foundation; Big Data; HDFS; Hadoop distributed file system; Linux; MapReduce; commodity hardware; open source cloud computing platform; software programming framework; Data handling; Data storage systems; Distributed databases; File systems; Hardware; Information management; Reliability; Big Data; Datanode; HDFS; Hadoop; MapReduce; Namenode;
Conference_Titel :
Advance Computing Conference (IACC), 2014 IEEE International
Conference_Location :
Gurgaon
Print_ISBN :
978-1-4799-2571-1
DOI :
10.1109/IAdCC.2014.6779407