🤖 AI Summary
This work addresses the tendency of vision-language models to generate “commonsense-driven hallucinations”—overreliance on prior knowledge at the expense of visual evidence—when confronted with conflicting image content. The study formally defines and quantifies this phenomenon, introducing CDH-Bench, a novel benchmark comprising anomalous scenarios across counting, relational, and attribute-based tasks. It proposes a multidimensional conflict testing framework alongside new evaluation metrics, including counterfactual accuracy and commonsense collapse rate. Through binary classification and multiple-choice question-answering tasks, the paper systematically evaluates the visual fidelity of state-of-the-art models. Experimental results reveal a pervasive failure among current models to prioritize visual input over ingrained priors, thereby demonstrating the necessity and effectiveness of the proposed benchmark for diagnosing and advancing model robustness.
📝 Abstract
Vision-language models (VLMs) achieve strong performance on many benchmarks, yet a basic reliability question remains underexplored: when visual evidence conflicts with commonsense, do models follow what is shown or what commonsense suggests? A characteristic failure in this setting is that the model overrides visual evidence and outputs the commonsense alternative. We term this phenomenon \textbf{commonsense-driven hallucination} (CDH). To evaluate it, we introduce \textbf{CDH-Bench}, a benchmark designed to create explicit \textbf{visual evidence--commonsense conflicts}. CDH-Bench covers three dimensions: \textit{counting anomalies}, \textit{relational anomalies}, and \textit{attribute anomalies}. We evaluate frontier VLMs under \textit{binary Question Answering (QA)} and \textit{multiple-choice QA}, and report metrics including \textit{Counterfactual Accuracy} (CF-Acc), \textit{Commonsense Accuracy} (CS-Acc), \textit{Counterfactual Accuracy Drop} (CFAD), \textit{Commonsense Collapse Rate} (CCR), and \textit{Relative Prior Dependency} (RPD). Results show that even strong models remain vulnerable to prior-driven normalization under visual evidence--commonsense conflict. CDH-Bench provides a controlled diagnostic of visual fidelity under visual evidence--commonsense conflict.