Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

MIT OpenCourseWare

Scaling Rules for Optimization - Lecture 7

MIT OpenCourseWare via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore neural computation from a spectral perspective in this 1 hour 21 minute lecture from MIT's Deep Learning course. Delve into the mathematical foundations of how neural networks process information through the lens of spectral analysis, examining the eigenvalue decomposition of network operations and their implications for learning dynamics. Investigate feature learning mechanisms and understand how networks automatically discover relevant representations from data. Master the principles of hyperparameter transfer, learning systematic approaches to adapt optimization settings across different network configurations. Discover scaling rules that govern how hyperparameters should be adjusted when transferring between networks of varying width and depth, providing practical guidelines for efficient model design and training. Gain insights into the theoretical underpinnings that connect network architecture choices to optimization behavior, enabling more informed decisions in deep learning system design.

Syllabus

Lec 07. Scaling Rules for Optimization

Taught by

MIT OpenCourseWare

Reviews

Start your review of Scaling Rules for Optimization - Lecture 7

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.