Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Quantized LLM Training at Scale with ZeRO++

MLOps.community via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn how to dramatically reduce large language model training time through advanced quantization techniques in this conference talk from AI in Production 2025. Discover ZeRO++, a groundbreaking approach that quantizes both weights and gradients during training to achieve a 4x reduction in communication volume, resulting in over 50% faster end-to-end training times. Explore the technical implementation details of this Microsoft DeepSpeed innovation that addresses communication bottlenecks in large-scale LLM training. Gain insights from Guanhua Wang, Senior Researcher on the DeepSpeed team at Microsoft, who led the ZeRO++ project and contributed to Microsoft Phi-3 model training, as he shares practical strategies for optimizing distributed training workflows and reducing computational overhead in production environments.

Syllabus

Quantized LLM Training at Scale with ZeRO++ // Guanhua Wang // AI in Production 2025

Taught by

MLOps.community

Reviews

Start your review of Quantized LLM Training at Scale with ZeRO++

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.