Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

Udemy

Data Engineering Project SQL, Python, Airflow, Docker, CI/CD

via Udemy

Overview

Become a Data Engineer by Learning APIs, SQL, Python, Docker, Airflow, CI/CD, Functional/ Data Quality Tests and more!

What you'll learn:
  • Build Python scripts for data extraction by interacting with APIs using Postman, loading into the data warehouse and transforming (ELT)
  • Use PostgreSQL as a data warehouse. Interact with the data warehouse using both psql & DBeaver
  • Discover how to containerize data applications using Docker, making your data pipelines portable and easy to scale.
  • Master the basics of orchestrating and automating your data workflows with Apache Airflow, a must-have tool in data engineering.
  • Understand how to perform unit, integration & end-to-end (E2E) tests using a combination of pytest and Airflow's DAG tests to validate your data pipelines.
  • Implement data quality tests using SODA to ensure your data meets business and technical requirements.
  • Learn to automate deployment pipelines using GitHub Actions to ensure smooth, continuous integration and delivery.

Data Engineering is the backbone of modern data-driven companies. To excel, you need experience with the tools and processes that power data pipelines in real-world environments. This course gives you practical, project-based learning with the following tools PostgreSQL, Python, Docker, Airflow, Postman, SODA and Github Actions. I will guide you as to how you can use these tools.


What you will learn in the course:


  1. Python for Data Engineering: Build Python scripts for data extraction by interacting with APIs using Postman, loading into the data warehouse and transforming (ELT)

  2. SQL for Data Pipelines: Use PostgreSQL as a data warehouse. Interact with the data warehouse using both psql & DBeaver

  3. Docker for Containerized Deployments: Discover how to containerize data applications using Docker, making your data pipelines portable and easy to scale.

  4. Airflow for Workflow Automation: Master the basics of orchestrating and automating your data workflows with Apache Airflow, a must-have tool in data engineering.

  5. Testing and Data Quality Assurance: Understand how to perform unit, integration & end-to-end (E2E) tests using a combination of pytest and Airflow's DAG tests to validate your data pipelines. Implement data quality tests using SODA to ensure your data meets business and technical requirements.

  6. CI/CD for Automated Testing & Deployment: Learn to automate deployment pipelines using GitHub Actions to ensure smooth, continuous integration and delivery.

Syllabus

  • Introduction
  • Data Extraction using API
  • Docker
  • Airflow
  • Postgres Data Warehouse
  • Testing
  • CI/CD

Taught by

Matthew Schembri

Reviews

4.5 rating at Udemy based on 362 ratings

Start your review of Data Engineering Project SQL, Python, Airflow, Docker, CI/CD

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.