Cloudera Administrator Training at Pune (25-28 Sept 2014)

Cloudera Administrator Training at Pune (25-28 Sept 2014)

 

  • Non Certification Training

    Sale Date Ended

    INR 62000
    Sold Out
  • Certification Training

    Sale Date Ended

    INR 74400
    Sold Out

Invite friends

Contact Us

Page Views : 567

About The Event

EARLY BIRD DISCOUNT CODE - CA2509

This four-day administrator training course for Apache Hadoop provides a comprehensive understanding of all the steps necessary to operate and maintain Hadoop clusters.

From installation and configuration, through load balancing and tuning your cluster, this Administration course has you covered.

Xebia is an official training partner of Cloudera, the leader in Apache Hadoop-based software and services.

Please note, that you need to bring your own laptop for this training.


Programme and Course Overview

Xebia is an official training partner of Cloudera, the leader in Apache Hadoop-based software and services.

Through lecture and interactive, hands-on exercises, this certified training will cover topics such as:

  • The internals of MapReduce and HDFS and how to build Hadoop architecture;
  • Proper cluster configuration and deployment to integrate with systems and hardware in the data center;
  • How to load data into the cluster from dynamically-generated files using Flume and from RDBMS using Sqoop;
  • Configuring the FairScheduler to provide service-level agreements for multiple users of a cluster;
  • Installing and implementing Kerberos-based security for your cluster;
  • Best practices for preparing and maintaining Apache Hadoop in production;
  • Troubleshooting, diagnosing, tuning and solving Hadoop issues.


Trainer's Profile
 

Tünde Bálint

Before joining Xebia, Tünde Balint worked as a technical consultant. As an architect she helped design and analyze the landscape of multiple international clients from different domains (telecom, oil, etc.). After seeing these landscapes she decided to focus on helping the customer to make data driven decisions. To achieve this she started to focus on Hadoop, Storm and algorithms which support data processing and analysis. 

Tünde joined Xebia on the 15th of May 2013 as a Big Data consultant.

Before becoming a consultant Tünde specialized in distributed computing while becoming a software engineer. She also worked at the National Institute of Subatomic Physics where she was focused on job submission to grid environments.

Gagan Agarwal

Gagan has a rich experience in domains like e-Governance, Document and Content Management, Customer Communication Managmement. Most of the projects he worked, involved huge data processing and were built on top of RDBMS, faced a lot of challenges in terms of scalability. Accordingly he started exploring other alternatives like Hadoop and Graph Databases and found them very interesting. He has mastered these technologies and applied them in various use cases. Gagan is also an active blogger/speaker and have been speaking on these technologies in conferences like IndicThreads 2012 and 2013.



Target Group & Prerequisites:

This course is best suited to system administrators and IT managers who have basic Linux experience.

Prior knowledge of Apache Hadoop is not required.


You Will Learn


• How the Hadoop Distributed File System and MapReduce work
• What hardware configurations are optimal for Hadoop clusters
• What network considerations to take into account when building out your cluster
• How to configure Hadoop's options for best cluster performance
• How to configure NameNode High Availability
• How to configure NameNode Federation
• How to configure the FairScheduler to provide service-level agreements for multiple users of a cluster
• How to install and implement Kerberos-based security for your cluster
• How to maintain and monitor your cluster
• How to load data into the cluster from dynamically-generated files using Flume and from relational database management systems using Sqoop
• What system administration issues exist with other Hadoop projects such as Hive, Pig, and HBase


Outline


• Introduction
• The Case for Apache Hadoop
• HDFS
• Getting Data into HDFS
• MapReduce
• Planning Your Hadoop Cluster
• Hadoop Installation and Initial Configuration
• Installing and Configuring Hive, Impala, and Pig
• Hadoop Clients
• Cloudera Manager
• Advanced Cluster Configuration
• Hadoop Security
• Managing and Scheduling Jobs
• Cluster Maintenance
• Cluster Monitoring and Troubleshooting
• Conclusion



Venue Map