Show simple item record

dc.contributor.authorChen, Yu-Hsin
dc.contributor.authorEmer, Joel S.
dc.contributor.authorSze, Vivienne
dc.date.accessioned2016-05-03T01:15:11Z
dc.date.available2016-05-03T01:15:11Z
dc.date.issued2016-06
dc.identifier.urihttp://hdl.handle.net/1721.1/102369
dc.description.abstractDeep convolutional neural networks (CNNs) are widely used in modern AI systems for their superior accuracy but at the cost of high computational complexity. The complexity comes from the need to simultaneously process hundreds of filters and channels in the high-dimensional convolutions, which involve a significant amount of data movement. Although highly-parallel compute paradigms, such as SIMD/SIMT, effectively address the computation requirement to achieve high throughput, energy consumption still remains high as data movement can be more expensive than computation. Accordingly, finding a dataflow that supports parallel processing with minimal data movement cost is crucial to achieving energy-efficient CNN processing without compromising accuracy. In this paper, we present a novel dataflow, called row-stationary (RS), that minimizes data movement energy consumption on a spatial architecture. This is realized by exploiting local data reuse of filter weights and feature map pixels, i.e., activations, in the high-dimensional convolutions, and minimizing data movement of partial sum accumulations. Unlike dataflows used in existing designs, which only reduce certain types of data movement, the proposed RS dataflow can adapt to different CNN shape configurations and reduces all types of data movement through maximally utilizing the processing engine (PE) local storage, direct inter-PE communication and spatial parallelism. To evaluate the energy efficiency of the different dataflows, we propose an analysis framework that compares energy cost under the same hardware area and processing parallelism constraints. Experiments using the CNN configurations of AlexNet show that the proposed RS dataflow is more energy efficient than existing dataflows in both convolutional (1.4x to 2.5x) and fully-connected layers (at least 1.3x for batch size larger than 16). The RS dataflow has also been demonstrated on a fabricated chip, which verifies our energy analysis.en_US
dc.language.isoen_US
dc.publisherInstitute of Electrical and Electronics Engineers (IEEE)en_US
dc.relation.isversionofhttp://isca2016.eecs.umich.edu/index.php/main-program/en_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourceSzeen_US
dc.titleEyeriss: A Spatial Architecture for Energy-Efficient Dataflow for Convolutional Neural Networksen_US
dc.typeArticleen_US
dc.identifier.citationChen, Yu-Hsin, Joel Emer, and Vivienne Sze. "Eyeriss: A Spatial Architecture for Energy-Efficient Dataflow for Convolutional Neural Networks." 43rd ACM/IEEE International Symposium on Computer Architecture (ISCA) (June 2016).en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Scienceen_US
dc.contributor.approverSze, Vivienneen_US
dc.contributor.mitauthorChen, Yu-Hsinen_US
dc.contributor.mitauthorSze, Vivienneen_US
dc.contributor.mitauthorEmer, Joel S.en_US
dc.relation.journalProceedings of the 43rd ACM/IEEE International Symposium on Computer Architecture (ISCA)en_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dspace.orderedauthorsChen, Yu-Hsin; Emer, Joel; Sze, Vivienneen_US
dc.identifier.orcidhttps://orcid.org/0000-0002-3459-5466
dc.identifier.orcidhttps://orcid.org/0000-0002-4403-956X
dc.identifier.orcidhttps://orcid.org/0000-0003-4841-3990
mit.licenseOPEN_ACCESS_POLICYen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record