A Study of Stochastic and Noisy Oracles in Unconstrained Continuous Optimization
Centre de recherches mathématiques - CRM via YouTube
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore the fundamental challenges and solutions in modern optimization algorithms through this mathematical lecture that examines how inexact computations affect algorithmic performance. Delve into the theoretical foundations of stochastic and noisy oracles in unconstrained continuous optimization, moving beyond classical methods like gradient descent that require exact gradient and function value computations. Learn about the practical necessity of working with approximations in real-world scenarios, including stochastic optimization where true quantities are expectations over distributions that can only be estimated through sample averages. Discover how derivative-free optimization approximates first-order derivatives using function values, and examine various other examples where randomization and extensions create inexact oracle scenarios. Gain insights into the classification system for different types of inexactness that emerge across various optimization settings, and understand how these different forms of approximation impact algorithmic behavior and convergence properties. Master the theoretical framework that underlies many contemporary optimization approaches used in machine learning, statistics, and computational mathematics where exact computations are either impossible or computationally prohibitive.
Syllabus
Katya Scheinberg: A study of stochastic and noisy oracles in unconstrained continuous optimization
Taught by
Centre de recherches mathématiques - CRM