Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Scaling Laws, Compute, and the Future of AI - Engineering Challenges Behind Training Frontier Models

Y Combinator via YouTube

Overview

Coursera Spring Sale
40% Off Coursera Plus Annual!
Grab it
Explore the engineering challenges and strategic decisions behind training frontier AI models like Claude in this comprehensive interview featuring Nick Joseph, Anthropic's Head of Pre-training, in conversation with Y Combinator General Partner Ankit Gupta. Discover the technical realities of managing thousands of GPUs, debugging complex infrastructure issues, and balancing computational resources between pre-training and reinforcement learning processes. Learn about scaling laws and how they create feedback loops between compute power, model capabilities, and revenue generation, while examining why infrastructure problems often present greater challenges than machine learning problems in AI development. Understand the evolution from various AI approaches to why next-word prediction became the dominant paradigm, and gain insights into building early infrastructure at Anthropic. Examine efficiency optimization techniques, debugging strategies at massive scale, and the composition of pre-training teams, including the balance between generalists and specialists. Delve into the complexities of distributed training across thousands of GPUs, compare working with different chip architectures including GPUs versus TPUs, and understand the relationship between pre-training and post-training processes like RLHF and reasoning models. Explore future considerations around data quality and availability challenges, and gain perspective on where pre-training technology is headed next, all while learning from Joseph's career journey from Vicarious to OpenAI to Anthropic.

Syllabus

00:00 – Introduction
01:05 – From Vicarious to OpenAI to Anthropic
06:40 – What pretraining is
11:20 – Why next-word prediction won out
16:05 – Scaling laws and the feedback loop of compute → models → revenue
21:50 – Building Anthropic’s early infrastructure
27:35 – Efficiency hacks and debugging at scale
33:10 – Generalists vs. specialists on the pretraining team
38:45 – Challenges of training across thousands of GPUs
44:15 – Working with new chips: GPUs vs. TPUs
49:00 – Pretraining vs. post-training RLHF and reasoning models
54:25 – The future of data quality and availability
59:10 – Where pretraining goes next
1:03:00 – Closing reflections

Taught by

Y Combinator

Reviews

Start your review of Scaling Laws, Compute, and the Future of AI - Engineering Challenges Behind Training Frontier Models

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.