Deep Augmentation: Self-Supervised Learning with Transformations in Activation Space

AmazUtah_NLP at SemEval-2024 Task 9: A MultiChoice Question Answering System for Commonsense Defying Reasoning


View a PDF of the paper titled Deep Augmentation: Self-Supervised Learning with Transformations in Activation Space, by Rickard Br”uel-Gabrielsson and 3 other authors

View PDF

Abstract:We introduce Deep Augmentation, an approach to implicit data augmentation using dropout or PCA to transform a targeted layer within a neural network to improve performance and generalization. We demonstrate Deep Augmentation through extensive experiments on contrastive learning tasks in NLP, computer vision, and graph learning. We observe substantial performance gains with Transformers, ResNets, and Graph Neural Networks as the underlying models in contrastive learning, but observe inverse effects on the corresponding supervised problems. Our analysis suggests that Deep Augmentation alleviates co-adaptation between layers, a problem exhibited by self-supervised learning where ground truth labels are not available. We use this observation to formulate a method for selecting which layer to target; in particular, our experimentation reveals that targeting deeper layers with Deep Augmentation outperforms augmenting the input data. The simple network- and modality-agnostic nature of this approach enables its integration into various machine learning pipelines.

Submission history

From: Rickard Brüel Gabrielsson [view email]
[v1]
Sat, 25 Mar 2023 19:03:57 UTC (1,470 KB)
[v2]
Mon, 26 Feb 2024 19:42:20 UTC (1,732 KB)
[v3]
Mon, 11 Nov 2024 15:49:16 UTC (2,289 KB)



Source link
lol

By stp2y

Leave a Reply

Your email address will not be published. Required fields are marked *

No widgets found. Go to Widget page and add the widget in Offcanvas Sidebar Widget Area.