(5 days)Overview
The main purpose of the course is to give students the ability plan and implement big data workflows on HDInsight.

*** NOTE: This class also available in On-Demand, eLearning Format ***
Too busy at work to miss 5 days out of the office to take this class? Consider the MOC On-Demand version of this course. Watch videos developed by Microsoft, take the same hands-on labs, access an online expert 24/7, and study at on your own time, at your own pace. For details on this alternative format, go toMOC On-Demand: 20775-Perform Data Engineering on Microsoft HDInsight

Prerequisites
In addition to their professional experience, students who attend this course should have:
 Programming experience using R, and familiarity with common R packages
 Knowledge of common statistical methods and data analysis best practices.
 Basic knowledge of the Microsoft Windows operating system and its core functionality.
 Working knowledge of relational databases.

Module 1: Getting Started with HDInsight
This module introduces Hadoop, the MapReduce paradigm, and HDInsight.
Lessons
 What is Big Data?
 Introduction to Hadoop
 Working with MapReduce Function
 Introducing HDInsight
Lab : Working with HDInsight
 Provision an HDInsight cluster and run MapReduce jobs

Module 2: Deploying HDInsight Clusters
This module provides an overview of the Microsoft Azure HDInsight cluster types, in addition to the creation and maintenance of the HDInsight clusters. The module also demonstrates how to customize clusters by using script actions through the Azure Portal, Azure PowerShell, and the Azure command-line interface (CLI). This module includes labs that provide the steps to deploy and manage the clusters.
Lessons
 Identifying HDInsight cluster types
 Managing HDInsight clusters by using the Azure portal
 Managing HDInsight Clusters by using Azure PowerShell
Lab : Managing HDInsight clusters with the Azure Portal
 Create an HDInsight cluster that uses Data Lake Store storage
 Customize HDInsight by using script actions
 Delete an HDInsight cluster

Module 3: Authorizing Users to Access Resources
This module provides an overview of non-domain and domain-joined Microsoft HDInsight clusters, in addition to the creation and configuration of domain-joined HDInsight clusters. The module also demonstrates how to manage domain-joined clusters using the Ambari management UI and the Ranger Admin UI. This module includes the labs that will provide the steps to create and manage domain-joined clusters.
Lessons
 Non-domain Joined clusters
 Configuring domain-joined HDInsight clusters
 Manage domain-joined HDInsight clusters
Lab : Authorizing Users to Access Resources
 Prepare the Lab Environment
 Manage a non-domain joined cluster

Module 4: Loading data into HDInsight
This module provides an introduction to loading data into Microsoft Azure Blob storage and Microsoft Azure Data Lake storage. At the end of this lesson, you will know how to use multiple tools to transfer data to an HDInsight cluster. You will also learn how to load and transform data to decrease your query run time.
Lessons
 Storing data for HDInsight processing
 Using data loading tools
 Maximising value from stored data
Lab : Loading Data into your Azure account
 Load data for use with HDInsight

Module 7: Design Batch ETL solutions for big data with Spark
This module provides an overview of Apache Spark, describing its main characteristics and key features. Before you start, its helpful to understand the basic architecture of Apache Spark and the different components that are available. The module also explains how to design batch Extract, Transform, Load (ETL) solutions for big data with Spark on HDInsight. The final lesson includes some guidelines to improve Spark performance.
Lessons
 What is Spark?
 ETL with Spark
 Spark performance
Lab : Design Batch ETL solutions for big data with Spark.
 Create a HDInsight Cluster with access to Data Lake Store
 Use HDInsight Spark cluster to analyze data in Data Lake Store
 Analyzing website logs using a custom library with Apache Spark cluster on HDInsight
 Managing resources for Apache Spark cluster on Azure HDInsight

Module 8: Analyze Data with Spark SQL
This module describes how to analyze data by using Spark SQL. In it, you will be able to explain the differences between RDD, Datasets and Dataframes, identify the uses cases between Iterative and Interactive queries, and describe best practices for Caching, Partitioning and Persistence. You will also look at how to use Apache Zeppelin and Jupyter notebooks, carry out exploratory data analysis, then submit Spark jobs remotely to a Spark cluster.
Lessons
 Implementing iterative and interactive queries
 Perform exploratory data analysis
Lab : Performing exploratory data analysis by using iterative and interactive queries
 Build a machine learning application
 Use zeppelin for interactive data analysis
 View and manage Spark sessions by using Livy

Module 9: Analyze Data with Hive and Phoenix
In this module, you will learn about running interactive queries using Interactive Hive (also known as Hive LLAP or Live Long and Process) and Apache Phoenix. You will also learn about the various aspects of running interactive queries using Apache Phoenix with HBase as the underlying query engine.
Lessons
 Implement interactive queries for big data with interactive hive.
 Perform exploratory data analysis by using Hive
 Perform interactive processing by using Apache Phoenix
Lab : Analyze data with Hive and Phoenix
 Implement interactive queries for big data with interactive Hive
 Perform exploratory data analysis by using Hive
 Perform interactive processing by using Apache Phoenix

Module 10: Stream Analytics
The Microsoft Azure Stream Analytics service has some built-in features and capabilities that make it as easy to use as a flexible stream processing service in the cloud. You will see that there are a number of advantages to using Stream Analytics for your streaming solutions, which you will discuss in more detail. You will also compare features of Stream Analytics to other services available within the Microsoft Azure HDInsight stack, such as Apache Storm. You will learn how to deploy a Stream Analytics job, connect it to the Microsoft Azure Event Hub to ingest real-time data, and execute a Stream Analytics query to gain low-latency insights. After that, you will learn how Stream Analytics jobs can be monitored when deployed and used in production settings.
Lessons
 Stream analytics
 Process streaming data from stream analytics
 Managing stream analytics jobs
Lab : Implement Stream Analytics
 Process streaming data with stream analytics
 Managing stream analytics jobs

Module 11: Implementing Streaming Solutions with Kafka and HBase
In this module, you will learn how to use Kafka to build streaming solutions. You will also see how to use Kafka to persist data to HDFS by using Apache HBase, and then query this data.
Lessons
 Building and Deploying a Kafka Cluster
 Publishing, Consuming, and Processing data using the Kafka Cluster
 Using HBase to store and Query Data
Lab : Implementing Streaming Solutions with Kafka and HBase
 Create a virtual network and gateway
 Create a storm cluster for Kafka
 Create a Kafka producer
 Create a streaming processor client topology
 Create a Power BI dashboard and streaming dataset
 Create an HBase cluster
 Create a streaming processor to write to HBase