Show simple item record

dc.contributor.authorMadry, A
dc.contributor.authorMakelov, A
dc.contributor.authorSchmidt, L
dc.contributor.authorTsipras, D
dc.contributor.authorVladu, A
dc.date.accessioned2021-11-05T14:56:05Z
dc.date.available2021-11-05T14:56:05Z
dc.date.issued2018
dc.identifier.urihttps://hdl.handle.net/1721.1/137496
dc.description.abstract© Learning Representations, ICLR 2018 - Conference Track Proceedings.All right reserved. Recent work has demonstrated that neural networks are vulnerable to adversarial examples, i.e., inputs that are almost indistinguishable from natural data and yet classified incorrectly by the network. To address this problem, we study the adversarial robustness of neural networks through the lens of robust optimization. This approach provides us with a broad and unifying view on much prior work on this topic. Its principled nature also enables us to identify methods for both training and attacking neural networks that are reliable and, in a certain sense, universal. In particular, they specify a concrete security guarantee that would protect against a well-defined class of adversaries. These methods let us train networks with significantly improved resistance to a wide range of adversarial attacks. They also suggest robustness against a first-order adversary as a natural security guarantee. We believe that robustness against such well-defined classes of adversaries is an important stepping stone towards fully resistant deep learning models.en_US
dc.language.isoen
dc.relation.isversionofhttps://openreview.net/forum?id=rJzIBfZAben_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourcearXiven_US
dc.titleTowards deep learning models resistant to adversarial attacksen_US
dc.typeArticleen_US
dc.identifier.citationMadry, A, Makelov, A, Schmidt, L, Tsipras, D and Vladu, A. 2018. "Towards deep learning models resistant to adversarial attacks." 6th International Conference on Learning Representations, ICLR 2018 - Conference Track Proceedings.
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
dc.contributor.departmentMassachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory
dc.relation.journal6th International Conference on Learning Representations, ICLR 2018 - Conference Track Proceedingsen_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2021-02-02T14:23:22Z
dspace.orderedauthorsMadry, A; Makelov, A; Schmidt, L; Tsipras, D; Vladu, Aen_US
dspace.date.submission2021-02-02T14:23:31Z
mit.licenseOPEN_ACCESS_POLICY
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record