Tightening Information-Theoretic Generalization Bounds with Data-Dependent Estimates with an Application to SGLD - Daniel Roy
Institute for Advanced Study via YouTube
UC San Diego Product Management Certificate — AI-Powered PM Training
Learn AI, Data Science & Business — Earn Certificates That Get You Hired
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Explore a workshop presentation on tightening information-theoretic generalization bounds with data-dependent estimates, focusing on their application to Stochastic Gradient Langevin Dynamics (SGLD). Delve into the nature of generalization understanding, open problems, and barriers in the field. Examine non-vacuous bounds, stochastic gradient dynamics, and expected generalization error. Learn from Daniel Roy of the University of Toronto as he discusses these advanced concepts in deep learning theory, providing insights into current challenges and potential future directions in the field.
Syllabus
Intro
The nature of generalization understanding
Open problem
Barriers
Non vacuous bounds
Stochastic gradient dynamics
Expected generalization error
Plot
Conclusion
Taught by
Institute for Advanced Study