View a PDF of the paper titled MoPE: Parameter-Efficient and Scalable Multimodal Fusion via Mixture of Prompt Experts, by Ruixiang Jiang and 2 other authors
Abstract:Despite the demonstrated parameter efficiency of prompt-based multimodal fusion methods, their limited adaptivity and expressiveness often result in suboptimal performance compared to other tuning approaches. In this paper, we address these limitations by decomposing the vanilla prompts to adaptively capture instance-level features. Building upon this decomposition, we introduce the mixture of prompt experts (MoPE) technique to enhance the expressiveness of prompt tuning. MoPE leverages multimodal pairing priors to route the most effective prompt on a per-instance basis. Compared to vanilla prompting, our MoPE-based fusion method exhibits greater expressiveness, scaling more effectively with the training data and the overall number of trainable parameters. We also investigate regularization terms for expert routing, which lead to emergent expert specialization during training, paving the way for interpretable soft prompting. Extensive experiments across six multimodal datasets spanning four modalities demonstrate that our method achieves state-of-the-art results for prompt fusion, matching or even surpassing the performance of fine-tuning while requiring only 0.8% of the trainable parameters. Code will be released: this https URL.
Submission history
From: Ruixiang Jiang [view email]
[v1]
Thu, 14 Mar 2024 17:47:10 UTC (8,790 KB)
[v2]
Wed, 11 Sep 2024 09:19:43 UTC (9,612 KB)
Source link
lol