Abstract
We revisit the general framework introduced by Fazylab et al. [SIAM J. Optim., 28 (2018), pp. 2654-2689] to construct Lyapunov functions for optimization algorithms in discrete and continuous time. For smooth, strongly convex objective functions, we relax the requirements necessary for such a construction. As a result, we are able to prove for Polyak's ordinary differential equations and for a two-parameter family of Nesterov algorithms rates of convergence that improve on those available in the literature. We analyze the interpretation of Nesterov algorithms as discretizations of the Polyak equation. We show that the algorithms are instances of additive Runge-Kutta integrators and discuss the reasons why most discretizations of the differential equation do not result in optimization algorithms with acceleration. We also introduce a modification of Polyak's equation and study its convergence properties. Finally, we extend the general framework to the stochastic scenario and consider an application to random algorithms with acceleration for overparameterized models; again we are able to prove convergence rates that improve on those in the literature.
| Original language | English |
|---|---|
| Pages (from-to) | 537-566 |
| Number of pages | 30 |
| Journal | SIAM Journal on Optimization |
| Volume | 35 |
| Issue number | 1 |
| Early online date | 12 Mar 2025 |
| DOIs | |
| Publication status | Published - Mar 2025 |
Keywords
- first order methods
- integral quadratic constraints
- linear matrix inequality
- Nesterov's accelerated method
- overparameterized models
- Polyak ODE
ASJC Scopus subject areas
- Software
- Theoretical Computer Science
- Applied Mathematics
Fingerprint
Dive into the research topics of 'On the Connections between Optimization Algorithms, Lyapunov Functions, and Differential Equations: Theory and Insights'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver