Cloudera Administrator Training l Bangalore | 17 - 20 Sep 2015

Cloudera Administrator Training l Bangalore | 17 - 20 Sep 2015

 

  • Administrator_ Pune

    Sale Date Ended

    INR 74400
    Sold Out
  • Administrator_ Pune - Certification Training

    Sale Date Ended

    INR 90400
    Sold Out

Invite friends

Contact Us

Page Views : 100

About The Event

Cloudera Administrator Training for Apache Hadoop 

This four-day administrator training course for Apache Hadoop provides a comprehensive understanding of all the steps necessary to operate and maintain Hadoop clusters.

From installation and configuration, through load balancing and tuning your cluster, this Administration course has you covered.

Xebia is an official training partner of Cloudera, the leader in Apache Hadoop-based software and services.

Please note, that you need to bring your own laptop for this training.


Programme and Course Overview

Xebia is an official training partner of Cloudera, the leader in Apache Hadoop-based software and services.

Through lecture and interactive, hands-on exercises, this certified training will cover topics such as:

 

• The internals of MapReduce and HDFS and how to build Hadoop architecture;

• Proper cluster configuration and deployment to integrate with systems and hardware in the data center;

• How to load data into the cluster from dynamically-generated files using Flume and from RDBMS using Sqoop;

• Configuring the FairScheduler to provide service-level agreements for multiple users of a cluster;

• Installing and implementing Kerberos-based security for your cluster;

• Best practices for preparing and maintaining Apache Hadoop in production;

• Troubleshooting, diagnosing, tuning and solving Hadoop issues.


Trainer's Profile
  

 

Bhavuk Chawla 

 

Bhavuk is a Professional Instructor & Software Architect with an exceptional history and experience of over 10 years in leading IT projects involving complex functional as well as non-functional requirements. Bhavuk has gotHadoop Instructor accreditations from Cloudera. Bhavuk worked in various domains including eCommerce, Insurance and Location Intelligence. Bhavuk has strong background in Designing, Architecture of Enterprise applications, Java, Spring, HDFS, Map Reduce, Pig, Hive, Impala, Hadoop Administration, Google Search Appliance, Demandware eCommerce Platform, Google Analytics, Maven, Enterprise Architect, Social Integration via OAuth etc. Bhavuk is actively involved in imparting trainings on various technologies including Google Search Appliance, HDFS, Map Reduce, Hive, Pig, Impala, Java, Software Craftsmanship etc.

Bhavuk has a strong expertise working on IaaS, PaaS and SaaS based environments. He has been consulting several CMM 5 organizations on best practice implementation, coding standard implementation, improve programming efficiency and code coverage, and implement automation testing and continuous integration processes.

Some of the automation tool, Bhavuk has been working & consulting on includes: Cruise Control, Jenkins, Hudson. Among the collaboration software, Bhavuk has an extensive experience with JIRA, both development and administration.

His corporate work experience includes, Technical Architect at Sapient Consulting Pvt. Limited and Solutions Architect at Pitney Bowes Software.

 

AREA OF TRAINING EXPERTISE:

  •  Software Craftsmanship – Software Design Principles (DRY, YAGNI, Solid Design Principles, Refactoring, Clean code), Software Design Patterns (Structured, Behaviour, Creational, Anti-patterns etc.), Test Driven Development (Junit, Mocks and Stubs, Pair Programming)
  •  Big Data technologies – HDFS, Map Reduce, Pig, Hive, Impala, Hadoop Administration
  • Code Analysis & Performance Optimization: Application health check, code & architecture health check, emphasis on heap and thread dump analysis, memory management for domain specific application.

 

Target Group & Prerequisites:

This course is best suited to system administrators and IT managers who have basic Linux experience. 

Prior knowledge of Apache Hadoop is not required.


You Will Learn


• How the Hadoop Distributed File System and MapReduce work
• What hardware configurations are optimal for Hadoop clusters
• What network considerations to take into account when building out your cluster
• How to configure Hadoop's options for best cluster performance
• How to configure NameNode High Availability
• How to configure NameNode Federation
• How to configure the FairScheduler to provide service-level agreements for multiple users of a cluster
• How to install and implement Kerberos-based security for your cluster
• How to maintain and monitor your cluster
• How to load data into the cluster from dynamically-generated files using Flume and from relational database management systems using Sqoop
• What system administration issues exist with other Hadoop projects such as Hive, Pig, and HBase


Outline


• Introduction
• The Case for Apache Hadoop
• HDFS
• Getting Data into HDFS
• MapReduce
• Planning Your Hadoop Cluster
• Hadoop Installation and Initial Configuration
• Installing and Configuring Hive, Impala, and Pig
• Hadoop Clients
• Cloudera Manager
• Advanced Cluster Configuration
• Hadoop Security
• Managing and Scheduling Jobs
• Cluster Maintenance
• Cluster Monitoring and Troubleshooting
• Conclusion


Swati BAgga | sbagga@xebia.com | 09582898322