InDistill: Information flow-preserving knowledge distillation for model compression

AmazUtah_NLP at SemEval-2024 Task 9: A MultiChoice Question Answering System for Commonsense Defying Reasoning


View a PDF of the paper titled InDistill: Information flow-preserving knowledge distillation for model compression, by Ioannis Sarridis and 4 other authors

View PDF
HTML (experimental)

Abstract:In this paper, we introduce InDistill, a method that serves as a warmup stage for enhancing Knowledge Distillation (KD) effectiveness. InDistill focuses on transferring critical information flow paths from a heavyweight teacher to a lightweight student. This is achieved via a training scheme based on curriculum learning that considers the distillation difficulty of each layer and the critical learning periods when the information flow paths are established. This procedure can lead to a student model that is better prepared to learn from the teacher. To ensure the applicability of InDistill across a wide range of teacher-student pairs, we also incorporate a pruning operation when there is a discrepancy in the width of the teacher and student layers. This pruning operation reduces the width of the teacher’s intermediate layers to match those of the student, allowing direct distillation without the need for an encoding stage. The proposed method is extensively evaluated using various pairs of teacher-student architectures on CIFAR-10, CIFAR-100, and ImageNet datasets demonstrating that preserving the information flow paths consistently increases the performance of the baseline KD approaches on both classification and retrieval settings. The code is available at this https URL.

Submission history

From: Ioannis Sarridis [view email]
[v1]
Fri, 20 May 2022 07:40:09 UTC (784 KB)
[v2]
Thu, 24 Nov 2022 12:46:14 UTC (137 KB)
[v3]
Fri, 16 Jun 2023 14:32:05 UTC (183 KB)
[v4]
Wed, 22 Jan 2025 09:06:08 UTC (1,373 KB)



Source link
lol

By stp2y

Leave a Reply

Your email address will not be published. Required fields are marked *

No widgets found. Go to Widget page and add the widget in Offcanvas Sidebar Widget Area.