Show simple item record

dc.contributor.authorWu, Yannan
dc.contributor.authorTsai, Po-An
dc.contributor.authorMuralidharan, Saurav
dc.contributor.authorParashar, Angshuman
dc.contributor.authorSze, Vivienne
dc.contributor.authorEmer, Joel
dc.date.accessioned2024-01-04T13:48:13Z
dc.date.available2024-01-04T13:48:13Z
dc.date.issued2023-10-28
dc.identifier.isbn979-8-4007-0329-4
dc.identifier.urihttps://hdl.handle.net/1721.1/153277
dc.description.abstractDue to complex interactions among various deep neural network (DNN) optimization techniques, modern DNNs can have weights and activations that are dense or sparse with diverse sparsity degrees. To offer a good trade-off between accuracy and hardware performance, an ideal DNN accelerator should have high flexibility to efficiently translate DNN sparsity into reductions in energy and/or latency without incurring significant complexity overhead. This paper introduces hierarchical structured sparsity (HSS), with the key insight that we can systematically represent diverse sparsity degrees by having them hierarchically composed from multiple simple sparsity patterns. As a result, HSS simplifies the underlying hardware since it only needs to support simple sparsity patterns; this significantly reduces the sparsity acceleration overhead, which improves efficiency. Motivated by such opportunities, we propose a simultaneously efficient and flexible accelerator, named HighLight, to accelerate DNNs that have diverse sparsity degrees (including dense). Due to the flexibility of HSS, different HSS patterns can be introduced to DNNs to meet different applications’ accuracy requirements. Compared to existing works, HighLight achieves a geomean of up to 6.4 × better energy-delay product (EDP) across workloads with diverse sparsity degrees, and always sits on the EDP-accuracy Pareto frontier for representative DNNs.en_US
dc.publisherACM|56th Annual IEEE/ACM International Symposium on Microarchitectureen_US
dc.relation.isversionofhttps://doi.org/10.1145/3613424.3623786en_US
dc.rightsCreative Commons Attributionen_US
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/en_US
dc.titleHighLight: Efficient and Flexible DNN Acceleration with Hierarchical Structured Sparsityen_US
dc.typeArticleen_US
dc.identifier.citationWu, Yannan, Tsai, Po-An, Muralidharan, Saurav, Parashar, Angshuman, Sze, Vivienne et al. 2023. "HighLight: Efficient and Flexible DNN Acceleration with Hierarchical Structured Sparsity."
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
dc.identifier.mitlicensePUBLISHER_CC
dc.identifier.mitlicensePUBLISHER_CC
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2024-01-01T08:48:35Z
dc.language.rfc3066en
dc.rights.holderThe author(s)
dspace.date.submission2024-01-01T08:48:36Z
mit.licensePUBLISHER_CC
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record