Near-optimal Closed-loop Method via Lyapunov Damping for Convex Optimization
Abstract: We introduce an autonomous system with closed-loop damping for first-order convex optimization. While, to this day, optimal rates of convergence are almost exclusively achieved by non-autonomous methods via open-loop damping (e.g., Nesterov's algorithm), we show that our system, featuring a closed-loop damping, exhibits a rate arbitrarily close to the optimal one. We do so by coupling the damping and the speed of convergence of the system via a well-chosen Lyapunov function. By discretizing our system we then derive an algorithm and present numerical experiments supporting our theoretical findings.
- Finite time stabilization of nonlinear oscillators subject to dry friction. In Nonsmooth Mechanics and Analysis, pages 289–304, 2006.
- An extension of the second order dynamical system that models Nesterov’s convex gradient method. Applied Mathematics & Optimization, 84:1687–1716, 2021.
- A second-order gradient-like dissipative dynamical system with Hessian-driven damping: Application to optimization and mechanics. Journal de mathématiques pures et appliquées, 81(8):747–779, 2002.
- Asymptotic stabilization of inertial gradient dynamics with time-dependent viscosity. Journal of Differential Equations, 263(9):5412–5458, 2017.
- From the Ravine method to the Nesterov method and vice versa: a dynamical system perspective. SIAM Journal on Optimization, 32(3):2074–2101, 2022.
- Fast convergence of inertial dynamics and algorithms with asymptotic vanishing damping. Mathematical Programming, 168:123–175, 2015.
- A dynamic approach to a proximal-Newton method for monotone inclusions in Hilbert spaces, with complexity O(1/n2)𝑂1superscript𝑛2{O}(1/n^{2})italic_O ( 1 / italic_n start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT ). Journal of Convex Analysis, 23:139–180, 2016a.
- Fast convex optimization via inertial dynamics with Hessian driven damping. Journal of Differential Equations, 261(10):5734–5783, 2016b.
- Rate of convergence of the Nesterov accelerated gradient method in the subcritical case α≤𝛼absent\alpha\leqitalic_α ≤ 3. ESAIM: Control, Optimisation and Calculus of Variations, 25:2, 2019.
- Fast optimization via inertial dynamics with closed-loop damping. Journal of the European Mathematical Society, 25(5):1985–2056, 2022.
- Fast convex optimization via closed-loop time scaling of gradient dynamics. arXiv preprint arXiv:2301.00701, 2023.
- Optimal rate of convergence of an ODE associated to the fast gradient descent schemes for b>0𝑏0b>0italic_b > 0. HAL Preprint hal-01547251, 2017.
- Optimal convergence rates for Nesterov acceleration. SIAM Journal on Optimization, 29(4):3131–3153, 2019.
- Stochastic approximations and differential inclusions. SIAM Journal on Control and Optimization, 44(1):328–348, 2005.
- Subgradient methods. Lecture notes of EE392o, Stanford University, 2003.
- Asymptotics for some semilinear hyperbolic equations with non-autonomous damping. Journal of Differential Equations, 252(1):294–322, 2012.
- On the long time behavior of second order differential equations with asymptotically small dissipation. Transactions of the American Mathematical Society, 361(11):5983–6017, 2009.
- On the convergence of the iterates of the “fast iterative shrinkage/thresholding algorithm”. Journal of Optimization theory and Applications, 166:968–982, 2015.
- George Cybenko. Approximation by superpositions of a sigmoidal function. Mathematics of control, signals and systems, 2(4):303–314, 1989.
- David M Grobman. Homeomorphism of systems of differential equations. Doklady Akademii Nauk SSSR, 128(5):880–881, 1959.
- Philip Hartman. A lemma in the theory of structural stability of differential equations. Proceedings of the American Mathematical Society, 11(4):610–620, 1960.
- Asymptotics for a second-order differential equation with nonautonomous damping and an integrable source term. Applicable Analysis, 94(2):435–443, 2015.
- A control-theoretic perspective on optimal high-order optimization. Mathematical Programming, 195(1):929–975, 2022.
- Lennart Ljung. Analysis of recursive stochastic algorithms. IEEE transactions on automatic control, 22(4):551–575, 1977.
- Ramzi May. Long time behavior for a semilinear hyperbolic equation with asymptotically vanishing damping term and convex potential. Journal of Mathematical Analysis and Applications, 430(1):410–416, 2015.
- Ramzi May. Asymptotic for a second-order evolution equation with convex potential and vanishing damping term. Turkish Journal of Mathematics, 41(3):681–685, 2017.
- Yurii Nesterov. A method of solving a convex programming problem with convergence rate O(1k2)1superscript𝑘2\left(\frac{1}{k^{2}}\right)( divide start_ARG 1 end_ARG start_ARG italic_k start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT end_ARG ). In Doklady Akademii Nauk, volume 269(3), pages 543–547, 1983.
- Peter Philip. Ordinary differential equations, lecture notes, 2023. URL https://www.math.lmu.de/~philip/publications/lectureNotes/philipPeter_ODE.pdf. [Online; accessed 17.07.2023].
- Boris T. Polyak. Minimization of unsmooth functionals. USSR Computational Mathematics and Mathematical Physics, 9(3):14–29, 1969.
- B.T. Polyak. Some methods of speeding up the convergence of iteration methods. USSR Computational Mathematics and Mathematical Physics, 4(5):1–17, 1964.
- Evolutionary Equations: Picard’s Theorem for Partial Differential Equations, and Applications. Springer Nature, 2022.
- Understanding the acceleration phenomenon via high-resolution differential equations. Mathematical Programming, 195(1):79–148, 2021.
- A differential equation for modeling Nesterov’s accelerated gradient method: Theory and insights. In Advances in Neural Information Processing Systems (NeurIPS), volume 27, pages 1–9. Curran Associates, Inc., 2014.
- A variational perspective on accelerated methods in optimization. Proceedings of the National Academy of Sciences, 113(47):E7351–E7358, 2016.
- A Lyapunov analysis of accelerated methods in optimization. Journal of Machine Learning Research, 22(1):5040–5073, 2021.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.