SP822: ETL Part 3: Production (AWS Databricks)
In this course data engineers optimize and automate Extract, Transform, Load (ETL) workloads using stream processing, job recovery strategies, and automation strategies like REST API integration. By the end of this course you will schedule highly optimized and robust ETL jobs, debugging problems along the way.
NOTE: This course is specific to the Databricks Unified Analytics Platform (based on Apache Spark™). While you might find it helpful for learning how to use Apache Spark in other environments, it does not teach you how to use Apache Spark in those environments.
3-6 hours, 75% hands-on
The course is a series of six self-paced lessons available in both Scala and Python. A final capstone project involves refactoring a batch ETL job to a streaming pipeline. In the process, students run the workload as a job and monitor it. Each lesson includes hands-on exercises.
This version of the course is intended to be run on AWS Databricks.
Note: This course will not run on Databricks Community Edition.
During this course learners
- Perform an ETL job on a streaming data source
- Parameterize a code base and manage task dependencies
- Submit and monitor jobs using the REST API or Command Line Interface
- Design and implement a job failure recovery strategy using the principle of idempotence
- Optimize ETL queries using compression and caching best practices with optimal hardware choices
- Course Overview and Setup
- Streaming ETL
- Runnable Notebooks
- Scheduling Jobs
- Job Failure
- ETL Optimizations
- Capstone Project
- Primary Audience: Data Engineers
- ETL Part 1 (optional, but strongly encouraged)
- ETL Part 2 (optional, but strongly encouraged)
- Please be sure to use a supported browser.
This self-paced training course may be used by 1 user for 12 months from the date of purchase. It may not be transferred or shared with any other user.