HDPCD:Spark using Scala Udemy Free Download
What you'll learn:
- Learn Scala, Spark, HDFS etc for the preparation of HDPCD Spark certification
Requirements::
- Basic programming skills
- Hortonworks Sandbox or valid account for IT Versity Big Data labs or any Hadoop clusters where Hadoop, Hive and Spark are well integrated.
- Minimum memory required based on the environment you are using with 64 bit operating system
Description:
Course cover the overall syllabus of HDPCD:Spark Certification.
-
Scala Fundamentals - Basic Scala programming required using REPL
-
Getting Started with Spark - Different setup options, setup process
-
Core Spark - Transformations and Actions to process the data
-
Data Frames and Spark SQL - Leverage SQL skills on top of Data Frames created from Hive tables or RDD
-
One week complementary lab access
-
Exercises - A set of self evaluated exercises to test skills for certification purpose
After the course one will gain enough confidence to give the certification and crack it.
All the demos are given on our state of the art Big Data cluster. You can avail one week complementary lab access by filling this form which is provided as part of the welcome message.
Who this course is for:
- Any one who want to prepare for HDPCD Spark Certification using Scala
Course Details:
-
18 hours on-demand video
-
2 downloadable resources
-
Full lifetime access
-
Access on mobile and TV
-
Certificate of completion