[Submitted on 3 Sep 2024]
View a PDF of the paper titled How to Determine the Preferred Image Distribution of a Black-Box Vision-Language Model?, by Saeid Asgari Taghanaki and 2 other authors
Abstract:Large foundation models have revolutionized the field, yet challenges remain in optimizing multi-modal models for specialized visual tasks. We propose a novel, generalizable methodology to identify preferred image distributions for black-box Vision-Language Models (VLMs) by measuring output consistency across varied input prompts. Applying this to different rendering types of 3D objects, we demonstrate its efficacy across various domains requiring precise interpretation of complex structures, with a focus on Computer-Aided Design (CAD) as an exemplar field. We further refine VLM outputs using in-context learning with human feedback, significantly enhancing explanation quality. To address the lack of benchmarks in specialized domains, we introduce CAD-VQA, a new dataset for evaluating VLMs on CAD-related visual question answering tasks. Our evaluation of state-of-the-art VLMs on CAD-VQA establishes baseline performance levels, providing a framework for advancing VLM capabilities in complex visual reasoning tasks across various fields requiring expert-level visual interpretation. We release the dataset and evaluation codes at url{this https URL}.
Submission history
From: Saeid Asgari Taghanaki [view email]
[v1]
Tue, 3 Sep 2024 19:26:13 UTC (2,800 KB)
Source link
lol