Cloudera Developer Training for Apache Spark and Hadoop | Bangalore |30 June - 03 July 16

Cloudera Developer Training for Apache Spark and Hadoop | Bangalore |30 June - 03 July 16


Invite friends

Contact Us

Page Views : 87

About The Event

Cloudera Developer Training for Spark and Hadoop I

Xebia's four-day hands-on training course delivers the key concepts and expertise participants need to ingest and process data on a Hadoop cluster using the most up-to-date tools and techniques. Employing Hadoop ecosystem projects such as Spark, Hive, Flume, Sqoop, and Impala, this training course is the best preparation for the real-world challenges faced by Hadoop developers. Participants learn to identify which tool is the right one to use in a given situation, and will gain hands-on experience in developing using those tools.

Bhavuk Chawla

Bhavuk has got Hadoop Developer, Hadoop Administrator and Data Analyst Instructor accreditations from Cloudera. Bhavuk worked in various domains including eCommerce, Insurance and Location Intelligence. Bhavuk has strong background in Designing, Architecture of Enterprise Big Data applications, Spark Core, Spark SQL, Spark Streaming, Spark Machine Learning, Map Reduce, HDFS, Pig, Hive, Impala, Sqoop, Flume, Kafka, HBase, Hadoop Administration, Microsoft Azure, Amazon EC2, Ansible, Java, Spring, Google Search Appliance, Demandware eCommerce Platform, Google Analytics, Maven, Enterprise Architect, Social Integration via OAuth etc. Bhavuk is actively involved in imparting trainings on various technologies including Spark, Map Reduce, HDFS, Hive, Pig, Impala, HBase, Kafka, Flume, Sqoop, Java, Software Craftsmanship etc. 

Learn how to import data into your Apache Hadoop cluster and process it with Spark, Hive, Flume, Sqoop, Impala, and other Hadoop ecosystem tools


Through instructor-led discussion and interactive, hands-on exercises, participants will learn Apache Spark and how it integrates with the entire Hadoop ecosystem, learning:

  • How data is distributed, stored, and processed in a Hadoop cluster
  • How to use Sqoop and Flume to ingest data
  • How to process distributed data with Apache Spark
  • How to model structured data as tables in Impala and Hive
  • How to choose the best data storage format for different data usage patterns
  • Best practices for data storage

Audience and Prerequisites

This course is designed for developers and engineers who have programming experience. Apache Spark examples and hands-on exercises are presented in Scala and Python, so the ability to program in one of those languages is required. Basic familiarity with the Linux command line is assumed. Basic knowledge of SQL is helpful. Prior knowledge of Hadoop is not required.

CCA: Spark and Hadoop Developer Certification

CCA175 is a hands-on, practical exam using Cloudera technologies. Each user is given their own CDH5 (currently 5.3.2) cluster pre-loaded with Spark, Impala, Crunch, Hive, Pig, Sqoop, Kafka, Flume, Kite, Hue, Oozie, DataFu, and many others. In addition the cluster also comes with Python (2.6 and 3.4), Perl 5.10, Elephant Bird, Cascading 2.6, Brickhouse, Hive Swarm, Scala 2.11, Scalding, IDEA, Sublime, Eclipse, and NetBeans.

Cost:- INR 74,400+ 15% Service Tax

Contact:- Swati Bagga | | 0958 289 8322

Venue Map