Show simple item record

dc.contributor.authorWensing, Patrick M
dc.contributor.authorSlotine, Jean-Jacques
dc.date.accessioned2022-01-24T19:05:18Z
dc.date.available2022-01-24T19:05:18Z
dc.date.issued2020
dc.identifier.urihttps://hdl.handle.net/1721.1/139673
dc.description.abstract© 2020 Wensing, Slotine. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. This paper considers the analysis of continuous time gradient-based optimization algorithms through the lens of nonlinear contraction theory. It demonstrates that in the case of a timeinvariant objective, most elementary results on gradient descent based on convexity can be replaced by much more general results based on contraction. In particular, gradient descent converges to a unique equilibrium if its dynamics are contracting in any metric, with convexity of the cost corresponding to the special case of contraction in the identity metric. More broadly, contraction analysis provides new insights for the case of geodesically-convex optimization, wherein non-convex problems in Euclidean space can be transformed to convex ones posed over a Riemannian manifold. In this case, natural gradient descent converges to a unique equilibrium if it is contracting in any metric, with geodesic convexity of the cost corresponding to contraction in the natural metric. New results using semi-contraction provide additional insights into the topology of the set of optimizers in the case when multiple optima exist. Furthermore, they show how semi-contraction may be combined with specific additional information to reach broad conclusions about a dynamical system. The contraction perspective also easily extends to time-varying optimization settings and allows one to recursively build large optimization structures out of simpler elements. Extensions to natural primal-dual optimization and game-theoretic contexts further illustrate the potential reach of these new perspectives.en_US
dc.language.isoen
dc.publisherPublic Library of Science (PLoS)en_US
dc.relation.isversionof10.1371/JOURNAL.PONE.0236661en_US
dc.rightsCreative Commons Attribution 4.0 International licenseen_US
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/en_US
dc.sourcePLoSen_US
dc.titleBeyond convexity—Contraction and global convergence of gradient descenten_US
dc.typeArticleen_US
dc.identifier.citationWensing, Patrick M and Slotine, Jean-Jacques. 2020. "Beyond convexity—Contraction and global convergence of gradient descent." PLoS ONE, 15 (8).
dc.contributor.departmentMassachusetts Institute of Technology. Department of Mechanical Engineering
dc.contributor.departmentMassachusetts Institute of Technology. Department of Brain and Cognitive Sciences
dc.contributor.departmentMassachusetts Institute of Technology. Nonlinear Systems Laboratory
dc.relation.journalPLoS ONEen_US
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2022-01-24T19:00:07Z
dspace.orderedauthorsWensing, PM; Slotine, J-Jen_US
dspace.date.submission2022-01-24T19:00:09Z
mit.journal.volume15en_US
mit.journal.issue8en_US
mit.licensePUBLISHER_CC
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record