Get 35% Off CFI Certifications - Code CFI35
Google AI Professional Certificate - Learn AI Skills That Get You Hired
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore the spectral properties of nonlinear random matrices and their applications in high-dimensional statistical estimation and machine learning through this 17-minute conference talk. Discover how new classes of structured random matrices have emerged in modern statistical estimation and machine learning, particularly in understanding the training and generalization performance of large neural networks and the fundamental limits of high-dimensional signal recovery. Learn about the challenges posed by nonlinear transformations in these matrices, which introduce structural dependencies that traditional analysis techniques struggle to handle. Examine a set of equivalence principles that establish asymptotic connections between complex nonlinear random matrix ensembles and more tractable linear models. See how these principles can be applied to characterize the performance of kernel methods and random feature models across different scaling regimes, and gain insights into the in-context learning capabilities of attention-based Transformer networks. Understand the theoretical foundations that bridge classical random matrix theory with modern machine learning applications, providing a mathematical framework for analyzing the behavior of contemporary learning algorithms in high-dimensional settings.
Syllabus
Yue M Lu | Nonlinear Random Matrices in High-Dimensional Estimation and Learning
Taught by
Harvard CMSA