Gradient Descent on Infinitely Wide Neural Networks - Global Convergence and
International Mathematical Union via YouTube
Python, Prompt Engineering, Data Science — Build the Skills Employers Want Now
Free courses from frontend to fullstack and AI
Overview
Syllabus
Intro
Machine learning Scientific context
Parametric supervised machine learning
Convex optimization problems
Theoretical analysis of deep learning
Optimization for multi-layer neural networks
Gradient descent for a single hidden layer
Wasserstein gradient flow
Many particle limit and global convergence (Chizat and Bach, 2018)
From optimization to statistics
Interpolation regime
Logistic regression for two-layer neural networks
From RKHS norm to variation norm
Kernel regime
Optimizing over two layers
Comparison of kernel and feature learning regimes
Discussion
Conclusion
Taught by
International Mathematical Union