2. Understand the Basics

2. Understand the Basics

The Hortonworks Data Platform consists of three layers of components. A coordinated and tested
set of these components is sometimes referred to as the Stack.

Core Hadoop 1: The basic components of Apache Hadoop version
1.x.

Hadoop Distributed File System (HDFS) : A special purpose
file system designed to provide high-throughput access to data in a highly distributed
environment.

MapReduce: A framework for performing high volume distributed
data processing using the MapReduce programming paradigm.

Core Hadoop 2: The basic components of Apache Hadoop version
2.x.

Hadoop Distributed File System (HDFS): A special purpose
file system designed to provide high-throughput access to data in a highly distributed
environment.

YARN: A resource negotiator for managing high volume
distributed data processing. Previously part of the first version of MapReduce.

MapReduce 2 (MR2): A set of client libraries for computation
using the MapReduce programming paradigm and a History Server for logging job and task information.
Previously part of the first version of MapReduce.

Essential Hadoop: A set of Apache components designed to ease
working with Core Hadoop.

Apache Pig: A platform for creating higher level data flow
programs that can be compiled into sequences of MapReduce programs, using Pig Latin, the platform’s
native language.

Apache Hive: A tool for creating higher level SQL queries
using HiveQL, the tool’s native language, that can be compiled into sequences of MapReduce programs.
Included with Apache HCatalog.

Apache HCatalog: A metadata abstraction layer that insulates users
and scripts from how and where data is physically stored. Now part of Apache
Hive. Includes WebHCat, which provides a set of REST
APIs for HCatalog and related Hadoop components. Originally named Templeton.

Apache HBase: A distributed, column-oriented database that provides the ability to access and manipulate data randomly in the context of the large blocks that make up HDFS.