Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Scaling Large Language Models - Getting Started with Large-Scale Parallel Training of LLMs

MLOps World: Machine Learning in Production via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn to implement large-scale parallel training strategies for billion-parameter language models in this hands-on workshop. Explore fundamental parallelization techniques including data, tensor, and pipeline parallelism, and discover how to compose them effectively for training massive LLMs when single or few GPUs lack sufficient memory capacity. Master strategic data and parameter sharding across devices, efficient collective communication operations for synchronizing gradients and activations, and recent LLM-specific techniques such as context parallelism. Engage in live coding exercises and practical implementations to build each strategy from first principles, understand their trade-offs, and optimize communication patterns and memory usage for maximum training throughput across distributed hardware. Gain insights from an independent machine learning researcher with extensive experience advising startups and large companies, whose research has been cited nearly 2000 times and won awards including best paper at NeurIPS 2022.

Syllabus

Scaling Large Language Models: Getting Started with Large-Scale Parallel Training of LLMs

Taught by

MLOps World: Machine Learning in Production

Reviews

Start your review of Scaling Large Language Models - Getting Started with Large-Scale Parallel Training of LLMs

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.