View a PDF of the paper titled Preventing Collapse in Contrastive Learning with Orthonormal Prototypes (CLOP), by Huanran Li and 2 other authors
Abstract:Contrastive learning has emerged as a powerful method in deep learning, excelling at learning effective representations through contrasting samples from different distributions. However, neural collapse, where embeddings converge into a lower-dimensional space, poses a significant challenge, especially in semi-supervised and self-supervised setups. In this paper, we first theoretically analyze the effect of large learning rates on contrastive losses that solely rely on the cosine similarity metric, and derive a theoretical bound to mitigate this collapse. {Building on these insights, we propose CLOP, a novel semi-supervised loss function designed to prevent neural collapse by promoting the formation of orthogonal linear subspaces among class embeddings.} Unlike prior approaches that enforce a simplex ETF structure, CLOP focuses on subspace separation, leading to more distinguishable embeddings. Through extensive experiments on real and synthetic datasets, we demonstrate that CLOP enhances performance, providing greater stability across different learning rates and batch sizes.
Submission history
From: Huanran Li [view email]
[v1]
Wed, 27 Mar 2024 15:48:16 UTC (2,092 KB)
[v2]
Mon, 7 Oct 2024 16:07:23 UTC (1,246 KB)
Source link
lol