Scaling Laws, Compute, and the Future of AI - Engineering Challenges Behind Training Frontier Models

Scaling Laws, Compute, and the Future of AI - Engineering Challenges Behind Training Frontier Models

Y Combinator via YouTube Direct link

27:35 – Efficiency hacks and debugging at scale

7 of 14

7 of 14

27:35 – Efficiency hacks and debugging at scale

Class Central Classrooms beta

YouTube videos curated by Class Central.

Classroom Contents

Scaling Laws, Compute, and the Future of AI - Engineering Challenges Behind Training Frontier Models

Automatically move to the next video in the Classroom when playback concludes

  1. 1 00:00 – Introduction
  2. 2 01:05 – From Vicarious to OpenAI to Anthropic
  3. 3 06:40 – What pretraining is
  4. 4 11:20 – Why next-word prediction won out
  5. 5 16:05 – Scaling laws and the feedback loop of compute → models → revenue
  6. 6 21:50 – Building Anthropic’s early infrastructure
  7. 7 27:35 – Efficiency hacks and debugging at scale
  8. 8 33:10 – Generalists vs. specialists on the pretraining team
  9. 9 38:45 – Challenges of training across thousands of GPUs
  10. 10 44:15 – Working with new chips: GPUs vs. TPUs
  11. 11 49:00 – Pretraining vs. post-training RLHF and reasoning models
  12. 12 54:25 – The future of data quality and availability
  13. 13 59:10 – Where pretraining goes next
  14. 14 1:03:00 – Closing reflections

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.