Proximal Gradient Methods for Nonsmooth Nonconvex Minimax - A Unified Convergence Analysis
Erwin Schrödinger International Institute for Mathematics and Physics (ESI) via YouTube
Stuck in Tutorial Hell? Learn Backend Dev the Right Way
MIT Sloan: Lead AI Adoption Across Your Organization — Not Just Pilot It
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Explore a 26-minute conference talk on proximal gradient methods for nonsmooth nonconvex minimax problems, delivered at the "One World Optimization Seminar in Vienna" workshop held at the Erwin Schrödinger International Institute for Mathematics and Physics. Delve into the analysis of parallel and alternating proximal gradient schemes within a unified framework, expanding on general convergence mechanisms for nonconvex nonsmooth optimization. Discover pointwise global convergence results and refined complexity analyses, departing from the common focus on nearly approximate stationary solutions. Learn how this approach broadens the scope of addressable minimax problems through Non Euclidean proximal steps, extending convergence and complexity results to a wider setting. Gain insights from this joint work with Eyal Cohen, which advances beyond typical weakly convex/concave and smooth models in recent literature.
Syllabus
Marc Teboulle - Proximal Gradient Methods for Nonsmooth Nonconvex Minimax: A Unified Convergence...
Taught by
Erwin Schrödinger International Institute for Mathematics and Physics (ESI)