Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

Coursera

Optimize Spark Performance & Throughput

Coursera via Coursera

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
In large-scale data engineering environments, performance issues such as slow transformations, excessive shuffle operations, and unbalanced workloads can impact analytics, reporting, and SLA commitments. This course teaches you how to analyze, diagnose, and optimize Apache Spark applications so they run faster, more efficiently, and more reliably. In this course, you’ll start by learning the fundamentals of Spark job execution, including how stages, tasks, shuffle operations, and execution plans reveal where bottlenecks occur. You’ll explore Spark’s built-in monitoring tools to interpret job behavior. From there, you’ll apply practical optimization techniques, including improving data partitioning, mitigating data skew, optimizing joins, configuring caching strategies, and choosing efficient file formats. You’ll also learn how to tune executors, memory, cores, and dynamic allocation to balance cost and performance across workloads. Learners should be familiar with basic knowledge of Python and Spark DataFrames; familiarity with JSON and SQL. This course is designed for data engineers and developers who need to diagnose and optimize Spark jobs running on large-scale distributed data pipelines. By the end, you’ll have the skills to confidently apply advanced tuning strategies, improve throughput, reduce shuffle overhead, and optimize resource usage.

Syllabus

  • Analyzing Spark Job Execution & Metrics
    • This module introduces learners to Spark’s job execution model and key performance metrics. Learners will explore the Spark UI, interpret job stages, tasks, and shuffle metrics, and diagnose performance bottlenecks using real job logs.
  • Fixing Data Skew, Shuffle Issues & Inefficient Joins
    • This module teaches learners how to solve the most common Spark bottlenecks: data skew, excessive shuffling, inefficient joins, and poor partitioning. Learners apply practical techniques such as salting, repartitioning, broadcast joins, and AQE.
  • Tuning Executors, Memory & Parallelism to Meet SLAs
    • This module focuses on configuring Spark resources—executors, CPU, memory, dynamic allocation, parallelism—and tuning job parameters to maximize throughput and meet strict performance SLAs.

Taught by

Merna Elzahaby

Reviews

Start your review of Optimize Spark Performance & Throughput

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.