About the Course
We will cover all the components of the Hadoop ecosystem at a high-level and Spark in depth as Spark is gaining lot of momentum in the Big Data ecosystem due to its in-memory processing capabilities.
The course contains following five components
- Â Â Hadoop Ecosystem
- Â Â Â Python basics
- Â Â Â Spark to be covered in Detail
- Â Â Â Small Projects relevant to Spark
- Â Â Â Prepping up an individual for Job Market - Interview Questions and up to 2 mock interview
Â
Hadoop Ecosystem will be covered at a high-level
- Â Â Â Data
- Â Â Â Evolution of data
- Â Â Â Hadoop EcoSystem
- Â Â Â Hdfs
- Â Â Â MapReduce
- Â Â Â Yarn
- Â Â Â Hive
- Â Â Â H-base
- Â Â Â Sqoop
Python Basics
Python basics will be covered at a high-level
Spark will be covered in Detail
- Â Â Â Introduction to Spark
- Â Â Â Spark Toolset
- Â Â Â Structured API Overview
- Â Â Â Basic Structured Operations
- Â Â Â Working with Different Types of Data
- Â Â Â Aggregations
- Â Â Â Joins
- Â Â Â Data Source
- Â Â Â Spark SQL
- Â Â Â Datasets
- Â Â Â Resilient Distributed Datasets (RDDs)
- Â Â Â Distributed Shared Variables
- Â Â Â How Spark Runs on a Cluster
- Â Â Â Developing Spark Applications
- Â Â Â Monitoring and Debugging
Projects
We will provide a simple, medium and Complex projects to guide an individual to get a real time experience
Support:
- Â Â Â We shall provide support to candidates by providing them a interview questions.
- Â Â Â Providing the model resumes
- Â Â Â Taking up to 2 mock up interviews
Please call me for a demo class. You can take a call post demo class.