installation

NiFi introduction

NiFi will allows you to create various data pipelines in a very nice web GUI.
Inside NiFi, one event sent and handled by the system is called a flow file. Each event will be stored as file, containing attributes. Flow files will be received, transformed, routed, split, transferred by processors. Tons of processors are proposed by default, there are processors to:

Setting up a HDF cluster with Ambari

To have a fully functional cluster running HortonWorks Data Flow

Attention, read this first before starting the deployment of an HDF cluster

(Valid end of June 2017)
The last version of Ambari (2.5.1) is well supported on Ubuntu 16 LTS and Ubuntu 14 LTS. This is also the case for the full Hortonworks Data Platform stack (HDP, version 2.6.1). Besides being supported on Oracle Linux, Suse, CentOS, RedHat and Debian.

Setting up a HDP cluster with Ambari

To have a fully functional cluster running HortonWorks Data Platform

Presentation

The Apache Ambari project implements a Web GUI that can be used to help in provisioning, managing and monitoring an Apache Hadoop cluster. With the time, it has introduced support for many Open Source projects, part of the Hadoop ecosystem.
The Ambari server will enable you to:

We would like here not to explain in details how to install a Linux server from scratch, but useful tips to customize the installation once the base OS is installed.
Nowadays, installing Linux has became very straight forward, the modern installers just ask you to click next and you're on the go !