MIT Sloan: Lead AI Adoption Across Your Organization — Not Just Pilot It
Free courses from frontend to fullstack and AI
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Explore an advanced lecture on Principal Component Analysis (PCA) and spectral clustering presented by Kaizheng Wang at the ICBS2024 conference. Delve into a novel $\ell_p$ perturbation theory for a hollowed version of PCA in Hilbert spaces, designed to improve upon traditional PCA methods when dealing with heteroscedastic noises. Examine the entrywise behaviors of principal component score vectors and their approximation by linear functionals of the Gram matrix in $\ell_p$ norm. Investigate how the choice of $p$ affects optimal bounds in sub-Gaussian mixture models, leading to optimality guarantees for spectral clustering. Discover how this $\ell_p$ theory applies to contextual community detection, resulting in simple spectral algorithms that achieve the information threshold for exact recovery and optimal misclassification rates. Gain insights into this cutting-edge research that bridges statistical theory with practical applications in machine learning and data analysis over the course of 49 minutes.
Syllabus
Kaizheng Wang: An $\ell_{p}$ theory of PCA and spectral clustering #ICBS2024
Taught by
BIMSA