Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Neural Scaling for Small LMs and AI Agents - How Superposition Yields Robust Neural Scaling

Discover AI via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
This video presentation from MIT researchers explores the fundamental principles behind neural scaling laws, challenging the "bigger is better" paradigm in AI development. Discover how strong representation superposition enables AI models to use their capacity more efficiently, providing a geometric explanation for the consistent 1/m loss decay observed in language models. The 28-minute talk delves into why foundation models improve according to power-law relationships when scaled up, focusing on representation efficiency rather than just size. Based on the research paper "Superposition Yields Robust Neural Scaling" by Yizhou Liu, Ziming Liu, and Jeff Gore from Massachusetts Institute of Technology, this presentation offers valuable insights for understanding how even smaller language models and AI agents can achieve impressive performance through better information representation strategies.

Syllabus

Neural Scaling for Small LMs & AI Agents (MIT)

Taught by

Discover AI

Reviews

Start your review of Neural Scaling for Small LMs and AI Agents - How Superposition Yields Robust Neural Scaling

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.