Show simple item record

dc.date.accessioned2017-04-28T15:05:41Z
dc.date.available2017-04-28T15:05:41Z
dc.date.issued2017-07
dc.identifier.urihttp://hdl.handle.net/1721.1/108491
dc.description.abstractDeep convolutional neural networks (CNNs) are indispensable to state-of-the-art computer vision algorithms. However, they are still rarely deployed on battery-powered mobile devices, such as smartphones and wearable gadgets, where vision algorithms can enable many revolutionary real-world applications. The key limiting factor is the high energy consumption of CNN processing due to its high computational complexity. While there are many previous efforts that try to reduce the CNN model size or the amount of computation, we find that they do not necessarily result in lower energy consumption. Therefore, these targets do not serve as a good metric for energy cost estimation. To close the gap between CNN design and energy consumption optimization, we propose an energy-aware pruning algorithm for CNNs that directly uses the energy consumption of a CNN to guide the pruning process. The energy estimation methodology uses parameters extrapolated from actual hardware measurements. The proposed layer- by-layer pruning algorithm also prunes more aggressively than previously proposed pruning methods by minimizing the error in the output feature maps instead of the filter weights. For each layer, the weights are first pruned and then locally fine-tuned with aclosed-form least-square solution to quickly restore the accuracy. After all layers are pruned, the entire network is globally fine-tuned using back-propagation. With the proposed pruning method, the energy consumption of AlexNet and GoogLeNet is reduced by 3.7X and 1.6X, respectively, with less than 1% top-5 accuracy loss. We also show that reducing the number of target classes in AlexNet greatly decreases the number of weights, but has a limited impact on energy consumption.en_US
dc.language.isoen_US
dc.publisherIEEE Computer Societyen_US
dc.relation.isversionofhttp://cvpr2017.thecvf.com/en_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourceSzeen_US
dc.titleDesigning Energy-Efficient Convolutional Neural Networks using Energy-Aware Pruningen_US
dc.typeArticleen_US
dc.identifier.citationTien-Ju Yang et al. "Designing Energy-Efficient Convolutional Neural Networks using Energy-Aware Pruning." Conference on Computer Vision and Pattern Recognition (CVPR 2017), Hawaii Convention Center, July 21-26, Honolulu, Hawaii, USA.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Scienceen_US
dc.contributor.approverSze, Vivienneen_US
dc.contributor.mitauthorYang, Tien-Ju
dc.contributor.mitauthorChen, Yu-Hsin
dc.contributor.mitauthorSze, Vivienne
dc.relation.journalConference on Computer Vision and Pattern Recognition, CVPR 2017en_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dspace.orderedauthorsTien-Ju Yang, Yu-Hsin Chen, Vivienne Szeen_US
dspace.embargo.termsNen_US
dc.identifier.orcidhttps://orcid.org/0000-0003-4728-0321
dc.identifier.orcidhttps://orcid.org/0000-0002-4403-956X
dc.identifier.orcidhttps://orcid.org/0000-0003-4841-3990
dspace.mitauthor.errortrue
mit.licenseOPEN_ACCESS_POLICYen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record