[공지사항]을 빙자한 안부와 근황 
Show more

Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

When Seeing Overrides Knowing: Disentangling Knowledge Conflicts in Vision-Language Models

Created by
  • Haebom

Author

Francesco Ortu, Zhijing Jin, Diego Doimo, Alberto Cazzaniga

Outline

This paper addresses the phenomenon in which visual-language models (VLMs) encounter knowledge conflicts between internal parameter knowledge and external information when performing complex tasks using multiple knowledge sources. Such conflicts can lead to hallucinations and unreliable responses, but their working mechanisms are not yet known. In this paper, we introduce a multimodal counterfactual queries dataset that intentionally contradicts internal common-sense knowledge and analyze the mechanism by which VLMs resolve cross-modal conflicts. Using logit inspection, we identify a small number of heads that control conflicts, and show that these heads can be modified to induce the model to produce results based on internal knowledge or visual input. Finally, we show that the attention of these heads accurately identifies local regions that cause visual overrides, and that it is more accurate than gradient-based attribution.

Takeaways, Limitations

Takeaways:
Provides new insights into the knowledge conflict resolution mechanism of VLMs.
We present a method to identify specific heads that control collisions using logit tests.
This demonstrates that the output of a model can be controlled by manipulating the head.
Analysis of the attention mechanism clarifies the process by which visual information influences model output.
We present a visual region localization method that is more precise than gradient-based features.
Limitations:
Further research is needed to determine whether the proposed method is applicable to all VLMs or all types of knowledge conflicts.
Validation of generalization performance on multimodal semi-empirical query datasets is required.
Further analysis is needed on the interpretability of the logit test and head manipulation.
A more in-depth study is needed to determine how manipulation of a specific head affects other parts of the model.
👍