MIT Libraries logoDSpace@MIT

MIT
View Item 
  • DSpace@MIT Home
  • MIT Open Access Articles
  • MIT Open Access Articles
  • View Item
  • DSpace@MIT Home
  • MIT Open Access Articles
  • MIT Open Access Articles
  • View Item
JavaScript is disabled for your browser. Some features of this site may not work without it.

Acceleration by Stepsize Hedging: Multi-Step Descent and the Silver Stepsize Schedule

Author(s)
Altschuler, Jason; Parrilo, Pablo
Thumbnail
Download3708502.pdf (1.215Mb)
Publisher Policy

Publisher Policy

Article is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.

Terms of use
Article is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.
Metadata
Show full item record
Abstract
Can we accelerate the convergence of gradient descent without changing the algorithmÐjust by judiciously choosing stepsizes? Surprisingly, we show that the answer is yes. Our proposed Silver Stepsize Schedule optimizes strongly convex functions in � log� 2 ≈ � 0.7864 iterations, where � = 1 + √ 2 is the silver ratio and � is the condition number. This is intermediate between the textbook unaccelerated rate � and the accelerated rate � 1/2 due to Nesterov in 1983. The non-strongly convex setting is conceptually identical, and standard black-box reductions imply an analogous partially accelerated rate � − log� 2 ≈ � −0.7864 . We conjecture and provide partial evidence that these rates are optimal among all stepsize schedules. The Silver Stepsize Schedule is constructed recursively in a fully explicit way. It is non-monotonic, fractal-like, and approximately periodic of period � log� 2 . This leads to a phase transition in the convergence rate: initially super-exponential (acceleration regime), then exponential (saturation regime). The core algorithmic intuition is hedging between individually suboptimal strategiesÐshort steps and long stepsÐsince bad cases for the former are good cases for the latter, and vice versa. Properly combining these stepsizes yields faster convergence due to the misalignment of worst-case functions. The key challenge in proving this speedup is enforcing long-range consistency conditions along the algorithm’s trajectory. We do this by developing a technique that recursively glues constraints from diferent portions of the trajectory, thus removing a key stumbling block in previous analyses of optimization algorithms. More broadly, we believe that the concepts of hedging and multi-step descent have the potential to be powerful algorithmic paradigms in a variety of contexts in optimization and beyond. This paper publishes and extends the irst author’s 2018 Master’s Thesis (advised by the second author)Ðwhich established for the irst time that judiciously choosing stepsizes can enable acceleration in convex optimization. Prior to this thesis, the only such result was for the special case of quadratic optimization, due to Young in 1953.
Date issued
2024-12-13
URI
https://hdl.handle.net/1721.1/158132
Department
Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
Journal
Journal of the ACM
Publisher
ACM
Citation
Altschuler, Jason and Parrilo, Pablo. 2024. "Acceleration by Stepsize Hedging: Multi-Step Descent and the Silver Stepsize Schedule." Journal of the ACM.
Version: Final published version
ISSN
0004-5411

Collections
  • MIT Open Access Articles

Browse

All of DSpaceCommunities & CollectionsBy Issue DateAuthorsTitlesSubjectsThis CollectionBy Issue DateAuthorsTitlesSubjects

My Account

Login

Statistics

OA StatisticsStatistics by CountryStatistics by Department
MIT Libraries
PrivacyPermissionsAccessibilityContact us
MIT
Content created by the MIT Libraries, CC BY-NC unless otherwise noted. Notify us about copyright concerns.