[Submitted on 23 Dec 2024]
View a PDF of the paper titled Multimodal Learning with Uncertainty Quantification based on Discounted Belief Fusion, by Grigor Bezirganyan and 3 other authors
Abstract:Multimodal AI models are increasingly used in fields like healthcare, finance, and autonomous driving, where information is drawn from multiple sources or modalities such as images, texts, audios, videos. However, effectively managing uncertainty – arising from noise, insufficient evidence, or conflicts between modalities – is crucial for reliable decision-making. Current uncertainty-aware ML methods leveraging, for example, evidence averaging, or evidence accumulation underestimate uncertainties in high-conflict scenarios. Moreover, the state-of-the-art evidence averaging strategy struggles with non-associativity and fails to scale to multiple modalities. To address these challenges, we propose a novel multimodal learning method with order-invariant evidence fusion and introduce a conflict-based discounting mechanism that reallocates uncertain mass when unreliable modalities are detected. We provide both theoretical analysis and experimental validation, demonstrating that unlike the previous work, the proposed approach effectively distinguishes between conflicting and non-conflicting samples based on the provided uncertainty estimates, and outperforms the previous models in uncertainty-based conflict detection.
Submission history
From: Grigor Bezirganyan [view email]
[v1]
Mon, 23 Dec 2024 22:37:18 UTC (2,375 KB)
Source link
lol