Show simple item record

dc.contributor.advisorArvind Satyanarayan.en_US
dc.contributor.authorKherraz, Houssam.en_US
dc.contributor.otherMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science.en_US
dc.date.accessioned2020-09-15T21:56:46Z
dc.date.available2020-09-15T21:56:46Z
dc.date.copyright2020en_US
dc.date.issued2020en_US
dc.identifier.urihttps://hdl.handle.net/1721.1/127417
dc.descriptionThesis: M. Eng., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, May, 2020en_US
dc.descriptionCataloged from the official PDF of thesis.en_US
dc.descriptionIncludes bibliographical references (pages 55-57).en_US
dc.description.abstractWith growing concerns over how machine learning models behave in deployment, people in academia and industry are more interested than ever in gaining insights into the inner workings of these black-box models. Yet, the current toolbox to understand neural networks is limited. In this work, I propose a new tool, called the Neuron Activation Sorter (NAS), centered around a new paradigm in machine learning interpretability. This new framing aims to use dataset examples as the main interaction tool to learn about the model. The Neuron Activation Sorter (NAS) operates at different levels of granularity through two modes. The Individual Neuron mode operates at the neuron level, while the Layer Summary mode operates at the layer level. The Layer Summary mode shows the distribution of different classes over activation values for each neuron of a specific layer through a histogram of stacked charts. The Individual Neuron mode further explores that distribution by exposing all the dataset images in the histogram visually. Together, they provide intuition about both micro and macro behaviors. I explore how these tools can leverage dataset items to both intuitively draw conclusions on the inner workings of a model and form hypotheses on potential failures. I give concrete examples on the insights they provide by exploring two neural networks: a basic 5-layer Convolutional Neural Network trained on the Quickdraw dataset and a VGG-16 model trained on Imagenet. Both examples expose a taxonomy of neurons and particular insights that are hard to access through other tools like feature visualizations or saliency maps.en_US
dc.description.statementofresponsibilityby Houssam Kherraz.en_US
dc.format.extent57 pagesen_US
dc.language.isoengen_US
dc.publisherMassachusetts Institute of Technologyen_US
dc.rightsMIT theses may be protected by copyright. Please reuse MIT thesis content according to the MIT Libraries Permissions Policy, which is available through the URL provided.en_US
dc.rights.urihttp://dspace.mit.edu/handle/1721.1/7582en_US
dc.subjectElectrical Engineering and Computer Science.en_US
dc.titleLeveraging dataset examples for the interpretation of back-box deep learning modelsen_US
dc.typeThesisen_US
dc.description.degreeM. Eng.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Scienceen_US
dc.identifier.oclc1192561536en_US
dc.description.collectionM.Eng. Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Scienceen_US
dspace.imported2020-09-15T21:56:46Zen_US
mit.thesis.degreeMasteren_US
mit.thesis.departmentEECSen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record