Show simple item record

dc.contributor.advisorLalana Kagal.en_US
dc.contributor.authorTong, Schrasing.en_US
dc.contributor.otherMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science.en_US
dc.date.accessioned2020-09-15T22:07:14Z
dc.date.available2020-09-15T22:07:14Z
dc.date.copyright2020en_US
dc.date.issued2020en_US
dc.identifier.urihttps://hdl.handle.net/1721.1/127635
dc.descriptionThesis: S.M., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, May, 2020en_US
dc.descriptionCataloged from the official PDF of thesis.en_US
dc.descriptionIncludes bibliographical references (pages 67-69).en_US
dc.description.abstractMachine learning has risen greatly in popularity and many of its applications have begun to impact our daily life. However, recent research has shown that trained models could produce biased predictions and discriminate against socially vulnerable groups. To address these problems, researchers have proposed various definitions of fairness as desirable forms of equality, often relying on sensitive attributes such as race or gender. In the image domain, the study of bias exceeds beyond fairness and sensitive attributes as biased models tend to generalize poorly in real-world applications, especially when the distribution differs slightly from the test set. Detecting bias in these scenarios is extremely challenging due to the many possible causes of biases, a problem further exacerbated by the lack of explicit labels on these features.en_US
dc.description.abstractResearch on explanation generation focuses on providing insights on a model's decision-making process and claims to detect bias in image classification. In this thesis, I investigated whether this claim holds true by proposing a list of important characteristics, including the ability to detect the cause and degree of bias, efficiency in terms of human effort involved, human understandability, and scalability towards multiple biases, and analyzed whether two popular explanation mechanisms, namely GRAD-CAM and TCAV, actually achieve them. To this end, I curated two datasets with fine labels and balanced sample sizes for the biased features and trained models with different degrees of bias by altering the data composition. Doing so allowed me to generate explanations on different models and observe how explanations change as the underlying data bias, and in turn the set of discriminating features shift.en_US
dc.description.abstractAl- though explanations help detect bias in most scenarios, they produced noisy results and performed poorly in estimating the degree of bias present. Aside from assessing the limits of explanations in bias detection, the approach employed in this thesis also serves as a novel method to evaluate the faithfulness of generated explanations.en_US
dc.description.statementofresponsibilityby Schrasing Tong.en_US
dc.format.extent69 pagesen_US
dc.language.isoengen_US
dc.publisherMassachusetts Institute of Technologyen_US
dc.rightsMIT theses may be protected by copyright. Please reuse MIT thesis content according to the MIT Libraries Permissions Policy, which is available through the URL provided.en_US
dc.rights.urihttp://dspace.mit.edu/handle/1721.1/7582en_US
dc.subjectElectrical Engineering and Computer Science.en_US
dc.titleDetecting bias in image classification using model explanations/en_US
dc.typeThesisen_US
dc.description.degreeS.M.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Scienceen_US
dc.identifier.oclc1194646862en_US
dc.description.collectionS.M. Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Scienceen_US
dspace.imported2020-09-15T22:07:14Zen_US
mit.thesis.degreeMasteren_US
mit.thesis.departmentEECSen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record