This page organizes papers related to artificial intelligence published around the world. This page is summarized using Google Gemini and is operated on a non-profit basis. The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.
ViLBias: Detecting and Reasoning about Bias in Multimodal Content
Created by
Haebom
Author
Shaina Raza, Caesar Saleh, Azib Farooq, Emrul Hasan, Franklin Ogidi, Maximus Powers, Veronica Chatrath, Marcelo Lotif, Karanpal Sekhon, Roya Javadi, Haad Zahid, Anam Zahid, Vahid Reza Khazaie, Zhenyu Yu
Outline
This paper highlights the need for a model that detects bias in multimodal news by going beyond text classification and inferring text-image pairs. To this end, we present ViLBias, a VQA-style benchmark and framework. ViLBias uses a dataset of 40,945 text-image pairs collected from various news organizations, annotated with bias labels and concise justifications through a two-stage LLM-based annotation pipeline. We evaluate SLM, LLM, and VLM for closed-query classification and open-query inference (oVQA), and compare parameter-efficient tuning strategies. We demonstrate that integrating images with text improves detection accuracy, and LLM/VLM better captures subtle framing and text-image mismatches than SLM. A parameter-efficient method (LoRA/QLoRA/Adapters) recovers 97-99% of the overall fine-tuning performance with <5% of the learnable parameters. For oVQA, inference accuracy ranged from 52-79%, fidelity from 68-89%, and this was improved by instruction tuning. Closed query accuracy showed a strong correlation with inference accuracy. ViLBias provides a scalable benchmark and a robust baseline for multimodal bias detection and evidence quality.
Takeaways, Limitations
•
Takeaways:
◦
Using text and images together improves bias detection accuracy.
◦
LLM/VLM captures subtle biases better than SLM.
◦
Parameter-efficient tuning techniques can make models lighter without compromising performance.
◦
The model's inference ability can be evaluated through oVQA and improved through instruction tuning.
◦
There is a high correlation between closed query accuracy and inference ability.
•
Limitations:
◦
There is no Limitations specified in the paper itself. (For arXiv papers, the research may still be in progress, and Limitations may be added in a future version.)