AI, Data Science & Business Certificates from Google, IBM & Microsoft
Finance Certifications Goldman Sachs & Amazon Teams Trust
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Master the complete engineering process of building large language models from the ground up in this comprehensive 6-hour course using pure PyTorch. Begin with foundational transformer architecture concepts and progress through training a basic LLM, then advance to modern architectural improvements and scaling techniques. Explore cutting-edge approaches including Mixture-of-Experts (MoE) implementations, supervised fine-tuning methods, reward modeling frameworks, and reinforcement learning from human feedback using Proximal Policy Optimization (PPO). Gain hands-on experience with each stage of the LLM development lifecycle, from core transformer components through advanced alignment techniques, enabling you to build and customize your own language models with deep technical understanding of the underlying engineering principles.
Syllabus
0:00:00 Part 0 - Introduction
0:05:43 Part 1 - Core Transformer Architecture
0:40:24 Part 2 — Training a Tiny LLM
1:30:27 Part 3 — Modernizing the Architecture
2:33:53 Part 4 — Scaling Up
3:17:22 Part 5 — Mixture-of-Experts MoE
3:44:19 Part 6 — Supervised Fine-Tuning SFT
4:23:44 Part 7 — Reward Modeling
4:59:55 Part 8 — RLHF with PPO
Taught by
freeCodeCamp.org