Clientele ➞

Hadoop Developer Training

rsz_hadoop-olympian

Duration: 3 Days

Description

Gartner predicts that 4.4 Million Jobs will be created globally to support BigData. BigData is a popular term used to describe the exponential growth, availability and use of information, both structured and unstructured. It is imperative that organizations and IT leaders focus on the ever-increasing volume, variety and velocity of information that forms BigData.

Hadoop is the core platform for structuring BigData and solves the problem of making it useful for Analytics. Our course will teach all there is to learn about using Hadoop for BigData analysis and gives you a clear understanding about processing BigData with Hadoop.

Why learn about Processing BigData with Hadoop?

  • Businesses are now aware of the large volumes of data that they generate in their day to day transactions. They have also realized that this BigData can provide very valuable insights once analyzed
  • The massive volume of BigData and its unstructured format make it difficult to analyze BigData. Hadoop brings the ability to cheaply process large amounts of data, regardless of structure.
  • If you are an IT professional who wants to stay up to date with the current buzzword then this is the course for you.
  • Knowledge about processing BigData with Hadoop will also prove to be a huge Resume builder for Students who will be trying for Placements soon.
  • If you are a developer who is uncertain about how Hadoop works, this course will clear things up and save you lot of time and effort
  • If you are business that is planning to shift to Hadoop, then this is the right course for your employees to get trained.
  • Processing BigData with Hadoop will prove to be an answer to many questions at once.
  • The session will be handled by very experienced trainers who not only have immense knowledge but are also loaded with valuable experience

Objectives

  • This training is a beginners course on Hadoop. It covers setting up a Hadoop instance, cluster and installing Hadoop related projects. Development of MapReduce 1 and YARN applications. And an overview of Pig, Hive and HBase Projects.

Who should attend

  • A developer who wants to learn Hadoop but you don’t know where to start

 

  •  A team that is struggling to extract insights from large scale and fast growing data in traditional systems

 

  • A team that has decided to migrate from a RDBMS or a traditional data warehouse to Hadoop, but needs help getting started

Prerequisites

  • Basics of Java and Linux fundamentals

Course Outline

DAY 1

  • Understanding the Big Data Paradigm
  • Introduction to Apache Hadoop
  • Hadoop Architecture
  • Understanding the Hadoop Runtime Environment
  • Installing Hadoop
    • Exercise: Installing Hadoop using Scripts
  • Introduction to HDFS
    • Exercise: Setting up HDFS and adding Files to HDFS
    • Exercise: Hadoop FSShell commands
  • HDFS Internals
    • Exercise: Writing a HDFS Application in Java to read and write data

DAY 2

  • MapReduce fundamentals – Anatomy of a job, task. Handling Failures
    • Exercise: Writing your first MapReduce
  • MapReduce revisited
    • Exercise: Simple log analysis using MapReduce
  • Advanced MapReduce – Combiners and Partitioners, Scheduling, Failure
    • Exercise: Inverted Index MapReduce application using Combiners and Partitioners
  • MapReduce more concepts – Composite keys, comparators, Input/Output Formats, Sorting, Joins
  • Introduction to YARN
    • YARN Architecture – Resource Managers, Node Managers, Application Masters
    • Exercise: Yarn Application
  • Streaming Mapreduce

DAY 3

  • Exercise: Setting up a Hadoop Cluster using Vagrant, VirtualBox, Shell
  • Introduction to other Hadoop Projects
  • Sqoop, Hive, Pig, HBase, Oozie, Zookeeper
  • Exercise: Installing Hive
  • Exercise: Working with Hive
  • Exercise: Installing Pig
  • Exercise: Working with Pig

About The Trainer

Dr. Yash Mody

Hadoop, Big Data Solution Specialist, Adobe AEM Architect

Dr. Yyash-modyash Mody, PhD, has developed and architected several enterprise applications using platforms like Hadoop, Oracle ADF, SalesForce, IBM Websphere, Quartz, SAP, Adobe AEM, Adobe LiveCycle, Apache Flex, TIBCO etc. At CloudThat, he works with our customers like PwC, Fidelity, Western Union, GE, HP, Oracle, Mahindra Bristlecone, Flipkart, Aditi, Sonata etc. to help them understand various big data technologies and design solutions for modern usecases like social media analytics, web analytics etc.

Over the years, Yash has trained over 1500 developers and architects from over 30 organisations.

 

View LinkedIn Profile

Other Details

 

Questions?

For latest batch dates, fees, location and general inquiries, contact our sales team at: +91 8880002200 or email at sales@cloudthat.in

For purely technical queries about the course please contact Bhavesh at bhavesh@cloudthat.in

Upcoming Batches

TBA

Fill out my online form.
Recently Viewed Courses.
  • Hadoop Developer Training

  • Favorite Courses
    No Favourites added yet.

    Our Partners