Show simple item record

dc.contributor.authorZhang, Jingzhao
dc.contributor.authorMokhtari, Aryan
dc.contributor.authorSra, Suvrit
dc.contributor.authorJadbabaie-Moghadam, Ali
dc.date.accessioned2022-01-07T19:25:09Z
dc.date.available2021-11-04T16:18:35Z
dc.date.available2022-01-07T19:25:09Z
dc.date.issued2018
dc.identifier.urihttps://hdl.handle.net/1721.1/137357.2
dc.description.abstract© 2018 Curran Associates Inc..All rights reserved. We study gradient-based optimization methods obtained by directly discretizing a second-order ordinary differential equation (ODE) related to the continuous limit of Nesterov's accelerated gradient method. When the function is smooth enough, we show that acceleration can be achieved by a stable discretization of this ODE using standard Runge-Kutta integrators. Specifically, we prove that under Lipschitz-gradient, convexity and order-(s + 2) differentiability assumptions, the sequence of iterates generated by discretizing the proposed second-order ODE converges to the optimal solution at a rate of O(N−2 s+1 s ), where s is the order of the Runge-Kutta numerical integrator. Furthermore, we introduce a new local flatness condition on the objective, under which rates even faster than O(N−2) can be achieved with low-order integrators and only gradient information. Notably, this flatness condition is satisfied by several standard loss functions used in machine learning. We provide numerical experiments that verify the theoretical rates predicted by our results.en_US
dc.language.isoen
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.sourceNeural Information Processing Systems (NIPS)en_US
dc.titleDirect Runge-Kutta discretization achieves accelerationen_US
dc.typeArticleen_US
dc.identifier.citationZhang, Jingzhao, Mokhtari, Aryan, Sra, Suvrit and Jadbabaie, Ali. 2018. "Direct Runge-Kutta discretization achieves acceleration." Advances in Neural Information Processing Systems, 2018-December.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Laboratory for Information and Decision Systemsen_US
dc.contributor.departmentMassachusetts Institute of Technology. Institute for Data, Systems, and Societyen_US
dc.relation.journalAdvances in Neural Information Processing Systemsen_US
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2021-03-30T13:49:29Z
dspace.orderedauthorsZhang, J; Sra, S; Mokhtari, A; Jadbabaie, Aen_US
dspace.date.submission2021-03-30T13:49:30Z
mit.journal.volume2018-Decemberen_US
mit.licensePUBLISHER_POLICY
mit.metadata.statusPublication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record

VersionItemDateSummary

*Selected version