Iteratively Reweighted Kernel Machines Efficiently Learn Sparse Functions
Paul G. Allen School via YouTube
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn about iteratively reweighted kernel machines and their efficiency in learning sparse functions through this 14-minute workshop presentation by Libin Zhu from the University of Washington. Explore the mathematical foundations and practical applications of kernel-based learning algorithms specifically designed for sparse function approximation. Discover how iterative reweighting techniques can enhance the performance of kernel machines in handling high-dimensional data with sparse underlying structures. Gain insights into the theoretical guarantees and computational advantages of this approach, including convergence properties and sample complexity bounds. Understand the connections between sparsity-inducing regularization and kernel methods, and examine how these techniques can be applied to real-world machine learning problems where the target function exhibits sparse characteristics.
Syllabus
IFDS Workshop Short Talks–Iteratively Reweighted Kernel Machines Efficiently Learn Sparse Functions
Taught by
Paul G. Allen School