Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

MIT OpenCourseWare

Generalization Theory - Overparameterization, Double Descent, and Inductive Biases - Lecture 6

MIT OpenCourseWare via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore fundamental concepts in generalization theory for deep learning through this 81-minute lecture from MIT's Deep Learning course. Examine the phenomenon of overparameterization in neural networks and understand how models with more parameters than training examples can still generalize well. Investigate the double descent curve, which challenges traditional bias-variance tradeoffs by showing how test error can decrease again as model complexity increases beyond the interpolation threshold. Analyze the limitations of classical statistical learning theory, particularly VC dimension, in explaining the generalization capabilities of modern deep networks. Discover how inductive biases—the assumptions built into learning algorithms—play a crucial role in enabling deep learning models to generalize effectively from limited data. Learn why traditional generalization bounds often fail to explain the success of deep learning and explore alternative theoretical frameworks that better capture the generalization behavior of overparameterized models.

Syllabus

Lec 06. Generalization Theory

Taught by

MIT OpenCourseWare

Reviews

Start your review of Generalization Theory - Overparameterization, Double Descent, and Inductive Biases - Lecture 6

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.