Privacy-Preserving Data Deduplication for Enhancing Federated Learning of Language Models (Extended Version)

Every’s Master Plan


View a PDF of the paper titled Privacy-Preserving Data Deduplication for Enhancing Federated Learning of Language Models (Extended Version), by Aydin Abadi and 2 other authors

View PDF
HTML (experimental)

Abstract:Deduplication is a vital preprocessing step that enhances machine learning model performance and saves training time and energy. However, enhancing federated learning through deduplication poses challenges, especially regarding scalability and potential privacy violations if deduplication involves sharing all clients’ data. In this paper, we address the problem of deduplication in a federated setup by introducing a pioneering protocol, Efficient Privacy-Preserving Multi-Party Deduplication (EP-MPD). It efficiently removes duplicates from multiple clients’ datasets without compromising data privacy. EP-MPD is constructed in a modular fashion, utilizing two novel variants of the Private Set Intersection protocol. Our extensive experiments demonstrate the significant benefits of deduplication in federated learning of large language models. For instance, we observe up to 19.62% improvement in perplexity and up to 27.95% reduction in running time while varying the duplication level between 10% and 30%. EP-MPD effectively balances privacy and performance in federated learning, making it a valuable solution for large-scale applications.

Submission history

From: Vishnu Dasu [view email]
[v1]
Thu, 11 Jul 2024 03:10:27 UTC (1,779 KB)
[v2]
Wed, 4 Dec 2024 17:56:57 UTC (4,695 KB)



Source link
lol

By stp2y

Leave a Reply

Your email address will not be published. Required fields are marked *

No widgets found. Go to Widget page and add the widget in Offcanvas Sidebar Widget Area.