HDP Operations Hortonworks Data Flow (HW HDP OPS DF)

Course Overview

This course is designed for Data Stewards or Data Flow Managers who are looking forward to automate the flow of data between systems. Topics Include Introduction to NiFi, Installing and Configuring NiFi, Detail explanation of NiFi User Interface, Explanation of its components and Elements associated with each. How to Build a dataflow, NiFi Expression Language, Understanding NiFi Clustering, Data Provenance, Security around NiFi, Monitoring Tools and HDF Best practices.

3 days
    • Describe HDF, Apache NiFi and its use cases.
    • Describe NiFi Architecture
    • Understand Nifi Features and Characteristics.
    • Understand System requirements to run Nifi.
    • Understand Installing and Configuring NiFi
    • Understand NiFi user interface in depth.
    • Understand how to build a DataFlow using NiFi
    • Understand Processor and its Elements
    • Understand Connection and its Elements
    • Understand Processor Group and its elements
    • Understand Remote Processor Group and its Elements
    • Learn how to optimize a DataFlow
    • Learn how to use NiFi Expression language and its use.
    • Learn about Attributes and Templates in NiFi
    • Understand Concepts of NiFi Cluster
    • Explain Data Provenance in NiFi
    • Learn how to Secure NiFi
    • Learn How to effectively Monitor NiFi
    • Learn about HDF Best Practices
  • 50% Lecture/Discussion
    50% Hands on Labs

    • Manual Installation of NiFi
    • Building a WorkFLow
    • Working with Processor Groups
    • Working with Remote Processor Groups
    • Using NiFi Expression Language.
    • Understanding and using Templates.
    • Installing and Configuring NiFi Cluster
    • Securing NiFi
    • Monitoring NiFi
    • End Of the Course Project


    • Getting Familiar to NiFi User Interface
    • Anatomy of a Processor
    • Anatomy of a Connection
    • Data Provenance
  • Students should be familiar with programming principles and have previous experience in software development. Experience with Linux and a basic understanding of DataFlow tools would be helpful. No prior Hadoop experience required, but is very helpful.

  • Data Engineers, Integration Engineers and Architects who are looking to automate Data flow between systems.