Enhancing Self-Supervised Learning through Transformations in Higher Activation Space
Author(s)
Gabrielsson, Rickard Brüel
DownloadThesis PDF (982.9Kb)
Advisor
Solomon, Justin
Terms of use
Metadata
Show full item recordAbstract
We introduce Deep Augmentation, an approach to data augmentation using dropout to
dynamically transform a targeted layer within a neural network, with the option to use
the stop-gradient operation, offering significant improvements in model performance and
generalization. We demonstrate the efficacy of Deep Augmentation through extensive
experiments on contrastive learning tasks in computer vision and NLP domains, where we
observe substantial performance gains with ResNets and Transformers as the underlying
models. Our experimentation reveals that targeting deeper layers with Deep Augmentation
outperforms augmenting the input data, and the simple network- and data-agnostic nature of
this approach enables its seamless integration into computer vision and NLP pipelines.
Date issued
2023-09Department
Massachusetts Institute of Technology. Department of Electrical Engineering and Computer SciencePublisher
Massachusetts Institute of Technology