Nonlinear conjugate gradient methods: worst-case convergence rates via computer-assisted analyses
Author(s)
Das Gupta, Shuvomoy; Freund, Robert M.; Sun, Xu A.; Taylor, Adrien
Download10107_2024_2127_ReferencePDF.pdf (1.013Mb)
Open Access Policy
Open Access Policy
Creative Commons Attribution-Noncommercial-Share Alike
Terms of use
Metadata
Show full item recordAbstract
We propose a computer-assisted approach to the analysis of the worst-case convergence of nonlinear conjugate gradient methods (NCGMs). Those methods are known for their generally good empirical performances for large-scale optimization, while having relatively incomplete analyses. Using our computer-assisted approach, we establish novel complexity bounds for the Polak-Ribière-Polyak (PRP) and the Fletcher-Reeves (FR) NCGMs for smooth strongly convex minimization. In particular, we construct mathematical proofs that establish the first non-asymptotic convergence bound for FR (which is historically the first developed NCGM), and a much improved non-asymptotic convergence bound for PRP. Additionally, we provide simple adversarial examples on which these methods do not perform better than gradient descent with exact line search, leaving very little room for improvements on the same class of problems.
Date issued
2024-08-22Department
Massachusetts Institute of Technology. Operations Research Center; Sloan School of ManagementJournal
Mathematical Programming
Publisher
Springer Berlin Heidelberg
Citation
Das Gupta, S., Freund, R.M., Sun, X.A. et al. Nonlinear conjugate gradient methods: worst-case convergence rates via computer-assisted analyses. Math. Program. 213, 1–49 (2025).
Version: Author's final manuscript