The Mathematics of Large Machine Learning Models - Lecture 1
International Centre for Theoretical Sciences via YouTube
Learn the Skills Netflix, Meta, and Capital One Actually Hire For
Get 20% off all career paths from fullstack to AI
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Explore the mathematical foundations underlying large machine learning models in this lecture that challenges classical theoretical wisdom about AI. Discover why modern AI models succeed despite using non-convex optimization, high complexity that often memorizes training data, and architectures that rarely offer parsimonious representations of target distributions. Survey recent theoretical work addressing these conceptual challenges and understand how learning can occur in completely unexpected scenarios. Examine the disconnect between traditional machine learning theory that recommended convex optimization and controlled model complexity versus the reality of successful modern AI systems. Gain insights into the mathematical principles that govern large-scale machine learning models and the ongoing research efforts to bridge the gap between classical theory and contemporary AI practice.
Syllabus
Date & Time : 11 August 2025, 16:30 to 17:30
Taught by
International Centre for Theoretical Sciences