International Conference and Workshop on Communication, Computing and Virtualization |
Foundation of Computer Science USA |
ICWCCV2015 - Number 3 |
September 2015 |
Authors: Smita Chaturvedi, Nivedita Bhirud, Fiona Lowden |
ffeb689d-a56d-4093-aa39-1140f666bd0d |
Smita Chaturvedi, Nivedita Bhirud, Fiona Lowden . Solving Big Data Problem using Hadoop File System(HDFS). International Conference and Workshop on Communication, Computing and Virtualization. ICWCCV2015, 3 (September 2015), 0-0.
The data which is useful not only for one person but for all, that data is called as Big data or It's a data to be too big to be processed in a single machine is known as Big data. Big data are the data which are extremely large in size that may be analyses computationally to disclose the patterns, associations and trends etc. For Example: Many users visited the amazon site; in particular page for how many user visit that page, from which IP address they visit the page, for how many hours they visit the page etc information stored in the amazon site is known as the example of Big data. Huge amount of data is created by phone data, online stores and by research data. Potentially data is created fast, the data coming from different sources in various formats and not most data are worthless but some data does has low value. Hadoop solves the Big data problem using the concept HDFS (Hadoop Distributed File System). In this paper the running of map reduce code in apache Hadoop is shown. Hadoop solves the problem of Big data by storing the data in distributed form in different machines. There are plenty of data but that data have to be store in a cost effective way and process it efficiently.