Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Are you looking to build a career in data engineering? Or want to get hands-on experience with Snowflake for your next project? Whether you’re a student, early career professional, or an experienced data professional, this program will enable you to acquire the practical skills you’ll need to be competitive in the job market to land your next job, or to grow your existing career as a data professional.
Created and delivered by Snowflake’s own developer advocates, this program emphasizes hands-on learning with lots of in-product exercises to give you the confidence you’ll need to tackle real-life, on-the-job projects.
You will learn:
How to use Snowflake to ingest data at scale How to perform data transformations with SQL and Python How to deliver data products to visualize insights How to orchestrate data pipelines How to implement DevOps best practices for data pipelines How to implement the observability of data pipelines
Syllabus
- Course 1: Intro to Snowflake for Devs, Data Scientists, Data Engineers
- Course 2: Introduction to Modern Data Engineering with Snowflake
- Course 3: Advanced Data Engineering with Snowflake
Courses
-
This course introduces learners to Snowflake as a platform for building applications, data pipelines, and AI models and workflows. It takes them from zero Snowflake knowledge all the way to creating user-defined functions, using a Snowflake Cortex LLM function, editing a Streamlit app, and more. The course unfolds in three parts: First, participants learn to use Snowflake’s core objects such as virtual warehouses, stages, and databases. Then they learn about slightly more advanced objects and features such as time travel, cloning, user-defined functions, and stored procedures. Finally, they’re introduced to Snowflake’s capabilities for data engineering, generative AI, machine learning, and app development. Learners come away equipped to start building with Snowflake and to continue their Snowflake learning journeys. This course is a prerequisite for upcoming Snowflake courses on data engineering, AI, and apps.
-
This is a technical, hands-on course that teaches learners how to build modern and continuous data pipelines with Snowflake. It focuses specifically on the most practical Snowflake concepts and tools to get learners up and running quickly with building data pipelines. Learners start by learning about the "Ingestion-Transformation-Delivery" framework for modern data engineering, and dive deeper into each component of the framework by learning how to: - Ingest data into Snowflake at scale using a variety of powerful techniques - Perform data transformations with SQL or Snowpark - Extend data transformations with user-defined functions, stored procedures, streams, and Snowflake Dynamic Tables - Deliver valuable data products through Snowflake Marketplace, Streamlit in Snowflake, and Snowflake Native Applications - Orchestrate pipelines using tasks and DAGs Throughout the course, learners follow along with the instructor using a combination of Snowflake, Visual Studio Code, GitHub, and the command line. The course is supplemented with readings containing plenty of resources to level up the learner's understanding of specific concepts. Learners come away understanding how to build end-to-end, continuous data pipelines with Snowflake.
-
This is a technical, hands-on course that teaches you how to implement DevOps best practices to build data pipelines, and how to implement observability to maintain and monitor data pipeline health. The course focuses on the most practical Snowflake concepts, features, and tools to get you up and running quickly with these concepts. You'll start by learning about DevOps, DevOps practices, and how DevOps fits into the context of data engineering. You'll incorporate source control, declarative management of database objects, continuous delivery, and use a command-line interface to implement DevOps best practices into a data pipeline. You'll specifically learn how to: - Use Snowflake's git integration to add source control to your data pipeline - Use GitHub for team-wide collaboration on your data pipeline - Use CREATE OR ALTER to declaratively manage database objects - Use GitHub Actions to implement continuous delivery for your pipeline - Use Snowflake CLI to deploy changes into dedicated data environments You'll also learn about observability, and how to implement it to maintain and monitor the health and performance of your data pipeline. You'll specifically learn how to: - Use logs to keep a record of events that occur within your pipeline - Use traces to maintain a detailed journey of events for operations in your pipeline - Use alerts to monitor for specific conditions in your pipeline, and combine them with notifications to encourage action among team members if critical errors occur in the pipeline Throughout the course, you'll follow along with the instructor using a combination of Snowflake, Visual Studio Code, GitHub, and the command line. The course is supplemented with readings containing resources to level up your understanding of specific concepts. You'll come away understanding how to incorporate DevOps best practices into data pipelines, and how to use observability to monitor the health and performance of pipelines.
Taught by
Snowflake Northstar