15
Jul
[Submitted on 11 Jul 2024] View a PDF of the paper titled Automatic Pruning of Fine-tuning Datasets for Transformer-based Language Models, by Mohammadreza Tayaranian and 4 other authors View PDF Abstract:Transformer-based language models have shown state-of-the-art performance on a variety of natural language understanding tasks. To achieve this performance, these models are first pre-trained on general corpus and then fine-tuned on downstream tasks. Previous work studied the effect of pruning the training set of the downstream tasks on the performance of the model on its evaluation set. In this work, we propose an automatic dataset pruning method for the training…