[Submitted on 29 Aug 2024]
View a PDF of the paper titled GSTAM: Efficient Graph Distillation with Structural Attention-Matching, by Arash Rasti-Meymandi and 3 other authors
Abstract:Graph distillation has emerged as a solution for reducing large graph datasets to smaller, more manageable, and informative ones. Existing methods primarily target node classification, involve computationally intensive processes, and fail to capture the true distribution of the full graph dataset. To address these issues, we introduce Graph Distillation with Structural Attention Matching (GSTAM), a novel method for condensing graph classification datasets. GSTAM leverages the attention maps of GNNs to distill structural information from the original dataset into synthetic graphs. The structural attention-matching mechanism exploits the areas of the input graph that GNNs prioritize for classification, effectively distilling such information into the synthetic graphs and improving overall distillation performance. Comprehensive experiments demonstrate GSTAM’s superiority over existing methods, achieving 0.45% to 6.5% better performance in extreme condensation ratios, highlighting its potential use in advancing distillation for graph classification tasks (Code available at this https URL).
Submission history
From: Arash Rasti Meymandi [view email]
[v1]
Thu, 29 Aug 2024 19:40:04 UTC (1,153 KB)
Source link
lol