View a PDF of the paper titled Invariance Principle Meets Vicinal Risk Minimization, by Yaoyao Zhu and 4 other authors
Abstract:Deep learning models excel in computer vision tasks but often fail to generalize to out-of-distribution (OOD) domains. Invariant Risk Minimization (IRM) aims to address OOD generalization by learning domain-invariant features. However, IRM struggles with datasets exhibiting significant diversity shifts. While data augmentation methods like Mixup and Semantic Data Augmentation (SDA) enhance diversity, they risk over-augmentation and label instability. To address these challenges, we propose a domain-shared Semantic Data Augmentation (SDA) module, a novel implementation of Variance Risk Minimization (VRM) designed to enhance dataset diversity while maintaining label consistency. We further provide a Rademacher complexity analysis, establishing a tighter generalization error bound compared to baseline methods. Extensive evaluations on OOD benchmarks, including PACS, VLCS, OfficeHome, and TerraIncognita, demonstrate consistent performance improvements over state-of-the-art domain generalization methods.
Submission history
From: Yaoyao Zhu [view email]
[v1]
Mon, 8 Jul 2024 09:16:42 UTC (3,014 KB)
[v2]
Thu, 23 Jan 2025 15:42:16 UTC (4,378 KB)
Source link
lol