PERFORMANCE ANALYSIS OF MAPREDUCE WTH LARGE DATASETS USING HADOOP
2017
Big data [1] is a huge amount of data that cannot be managed by traditional data management systems. Hadoop [2] is a tool that is used to handle this big data. For storing and retrieving the bigdata hadoop distributed file[4] system(HDFS) and mapreduce [3] are used respectively. Even petabytes or terabytes of data can be stored and retrieved easily using these techniques. This paper provides introduction to hadoop HDFS and Mapreduce. In this paper we have used large datasets to analyse the performance of mapreduce technique. Number of bytes read and written while performing mapreduce task on input given is also observed. We have analysed the behaviour of mapreduce task by varying the amount of input given. Also the pattern of number of bytes read and written when given input is varied is also analysed.
Keywords:
- Correction
- Cite
- Save
- Machine Reading By IdeaReader
0
References
0
Citations
NaN
KQI