Show simple item record

dc.contributor.advisorSolomon, Justin
dc.contributor.advisorYurochkin, Mikhail
dc.contributor.advisorGreenewald, Kristjan
dc.contributor.authorJung, Luann
dc.date.accessioned2022-08-29T16:28:56Z
dc.date.available2022-08-29T16:28:56Z
dc.date.issued2022-05
dc.date.submitted2022-05-27T16:19:07.097Z
dc.identifier.urihttps://hdl.handle.net/1721.1/145039
dc.description.abstractTraditional machine learning methods such as empirical risk minimization (ERM) frequently encounter the issue of achieving high accuracy on average but low accuracy on certain subgroups, especially when there exist spurious correlations between the input data and label. Previous approaches for reducing the discrepancy between average and worst-group accuracies typically require expensive known subgroup annotations for either every training data point (as is the case in group distributionally robust optimization (DRO)), or every validation data point. Furthermore, these distributionally robust approaches tend to show reduced performance when outliers are also present in the data. Unfortunately, existing methods for improving subgroup performance cannot be simply combined with prior approaches for excluding outliers, as they often directly conflict. This work proposes a method for addressing both group robustness and outlier exclusion when training machine learning models that requires no previous knowledge about subgroups or outliers within the data. We focus on attempting to balance these two traditionally clashing goals by clustering the gradients of the losses with respect to the model parameters. In doing so, we find minority subgroups and exclude outliers under the assertion that gradients within groups behave similarly while outliers exhibit more randomized behavior. This work demonstrates an improvement to both average and worst-group accuracies compared to baselines and other previous methods when applied to the Waterbirds image dataset.
dc.publisherMassachusetts Institute of Technology
dc.rightsIn Copyright - Educational Use Permitted
dc.rightsCopyright MIT
dc.rights.urihttp://rightsstatements.org/page/InC-EDU/1.0/
dc.titleGradient Subgroup Scanning for Distributionally and Outlier Robust Models
dc.typeThesis
dc.description.degreeM.Eng.
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
mit.thesis.degreeMaster
thesis.degree.nameMaster of Engineering in Electrical Engineering and Computer Science


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record