GenQ: Quantization in Low Data Regimes with Generative Synthetic Data

AmazUtah_NLP at SemEval-2024 Task 9: A MultiChoice Question Answering System for Commonsense Defying Reasoning


View a PDF of the paper titled GenQ: Quantization in Low Data Regimes with Generative Synthetic Data, by Yuhang Li and 4 other authors

View PDF
HTML (experimental)

Abstract:In the realm of deep neural network deployment, low-bit quantization presents a promising avenue for enhancing computational efficiency. However, it often hinges on the availability of training data to mitigate quantization errors, a significant challenge when data availability is scarce or restricted due to privacy or copyright concerns. Addressing this, we introduce GenQ, a novel approach employing an advanced Generative AI model to generate photorealistic, high-resolution synthetic data, overcoming the limitations of traditional methods that struggle to accurately mimic complex objects in extensive datasets like ImageNet. Our methodology is underscored by two robust filtering mechanisms designed to ensure the synthetic data closely aligns with the intrinsic characteristics of the actual training data. In case of limited data availability, the actual data is used to guide the synthetic data generation process, enhancing fidelity through the inversion of learnable token embeddings. Through rigorous experimentation, GenQ establishes new benchmarks in data-free and data-scarce quantization, significantly outperforming existing methods in accuracy and efficiency, thereby setting a new standard for quantization in low data regimes. Code is released at url{this https URL}.

Submission history

From: Yuhang Li [view email]
[v1]
Thu, 7 Dec 2023 23:31:42 UTC (3,138 KB)
[v2]
Fri, 8 Mar 2024 22:15:22 UTC (4,698 KB)
[v3]
Tue, 17 Sep 2024 14:49:21 UTC (4,697 KB)



Source link
lol

By stp2y

Leave a Reply

Your email address will not be published. Required fields are marked *

No widgets found. Go to Widget page and add the widget in Offcanvas Sidebar Widget Area.