” I hear and I forget. I see and I remember. I do and I understand. ”- Confucius
Learn from Practitioners
Sessions are conducted by our Architects and Developers. Hands-on training based on real world project experiences. Internship opportunities, post training, to suitable candidates. Unlike other training programs, our sessions are conducted by our Architects and Developers. We don’t use dedicated trainers or power-point presentations.
Learn what is current
Big Data is a rapidly changing technology space. Techniques such as MapReduce, which were widely used few years back is now not prevalent. We focus on current technologies, that are used in big data projects today. We don’t spend time teaching outdated technologies like Java MapReduce and Pig. We focus on Spark, the primary tool used by industry today to solve BigData problems.
Get Certificates that matter
We help you clear industry recognised certification exams.
Technologies We Focus On
We cover Hadoop, MapReduce, HDFS, Pig, Hive, Spark, HBase, MongoDB, Cassandra, Flume, Kafka, etc., With this course, you will not only understand what these technologies are but also will get hands on experience by solving real world problems.
Sessions are conducted by our Architects and Developers. Hands-on training based on real world project experiences. Internship opportunities, post training,to suitable candidates. Unlike other training programs, our sessions are conducted by our Architects and Developers. We don’t use dedicated trainers or power-point presentations.
- Install and work with a real Hadoop installation right on your desktop with Hortonworks sandbox and Cloudera quick start VM
- Manage big data on a cluster with HDFS and process with MapReduce
- Design and develop programs to analyze data on Hadoop with Pig and Spark
- Store and query your data with Hive,MySQL,HBase, Cassandra, MongoDB, Drill, Phoenix, and Presto
- Learn managing cluster with YARN, Mesos, Zookeeper, Oozie, Zeppelin, and Hue
- Handling streaming data with Kafka, Flume
- Processing streaming data in real time with Spark Streaming, Flink, and Storm
This course covers over 25 different technologies in over 3 months oftraining. It’s filled with hands-on activities and exercises, so you get some real experience in using Hadoop – it’s not just theory.
You’ll find a range of activities in this course for people at every level. If you’re a project manager who just wants to learn the buzzwords, there are web UI’s for many of the activities in the course that require no programming knowledge. If you’re comfortable with command lines, we’ll show you how to work with them too. And if you’re a programmer, I’ll challenge you with writing real scripts on a Hadoop system using Scala, Pig Latin, and Python.
You’ll walk away from this course with a real, deep understanding of Hadoop and its associated distributed systems, and you can apply Hadoop to real-world problems.
What are the requirements?
You will need to bring laptop running 64-bit OS, if you want to participate in the hands-on activities and exercises. You must have at least 8GB of free RAM on your system; 10GB or more is recommended.
Some activities will require some prior programming experience, preferably in Python or Scala.
A basic familiarity with the SQL queries and Linux command line will be very helpful.
Who can attend training?
Software engineers and programmers who want to understand the larger Hadoop ecosystem, and use it to store, analyze, and vend “big data” at scale.
Project, program, or product managers who want to understand the lingo and high-level architecture of Hadoop.Data analysts and database administrators who are curious about Hadoop and how it relates to their work.
System architects who need to understand the components available in the Hadoop ecosystem, and how they fit together.