Show simple item record

dc.contributor.advisorMichael Carbin.en_US
dc.contributor.authorSiswanto, Arlene Elizabeth.en_US
dc.contributor.otherMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science.en_US
dc.date.accessioned2021-05-24T19:52:33Z
dc.date.available2021-05-24T19:52:33Z
dc.date.copyright2021en_US
dc.date.issued2021en_US
dc.identifier.urihttps://hdl.handle.net/1721.1/130708
dc.descriptionThesis: M. Eng., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, February, 2021en_US
dc.descriptionCataloged from the official PDF of thesis.en_US
dc.descriptionIncludes bibliographical references (pages 69-72).en_US
dc.description.abstractBlock sparsity imposes structural constraints on the weight patterns of sparse neural networks. The structure of sparsity has been shown to affect efficiency of sparse computation in the libraries, kernels, and hardware commonly used in machine learning. Much work in the pruning literature has focused on the unstructured pruning of individual weights, which has been shown to reduce the memory footprint of a network, but cannot achieve the computational speedups that have become increasingly coveted as neural networks become deeper and more complex. On the opposite end of granularity, neuron pruning and channel pruning are unable to reach the same level of sparsity as unstructured pruning without compromising accuracy. Block-sparse pruning is a middle ground between these two extremes, with the potential for pruning to greater sparsities while still being amenable for acceleration. Our fine-tuning experiments demonstrate that block-sparse pruning offers a tradeoff between granularity and accuracy; increasing block size results in a gradual decrease in accuracy. Our weight rewinding experiments show that increasing block size decreases the maximum sparsity obtainable when pruning a network early in training. Finally, we make the surprising observation that randomly reinitializing the pruned network structure results in the same accuracy regardless of block size.en_US
dc.description.statementofresponsibilityby Arlene Elizabeth Siswanto.en_US
dc.format.extent72 pagesen_US
dc.language.isoengen_US
dc.publisherMassachusetts Institute of Technologyen_US
dc.rightsMIT theses may be protected by copyright. Please reuse MIT thesis content according to the MIT Libraries Permissions Policy, which is available through the URL provided.en_US
dc.rights.urihttp://dspace.mit.edu/handle/1721.1/7582en_US
dc.subjectElectrical Engineering and Computer Science.en_US
dc.titleBlock sparsity and weight initialization in neural network pruningen_US
dc.typeThesisen_US
dc.description.degreeM. Eng.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Scienceen_US
dc.identifier.oclc1251801573en_US
dc.description.collectionM.Eng. Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Scienceen_US
dspace.imported2021-05-24T19:52:33Zen_US
mit.thesis.degreeMasteren_US
mit.thesis.departmentEECSen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record