Introduction To Apache Hadoop Eco System

In the last article we understood what is Bigdata? In this article we will see a big data framework called “Hadoop”. Hadoop is a software library which will enable the users to distribute and process the large amount of data using clusters of commodity servers. This project includes,

Hadoop Common – Contains common utilities to work with Hadoop

Hadoop Distributed File System(HDFS) – A distributed file system which will provide high throughput access to the data.