Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Robust Adaptation of Large Multimodal Models for Retrieval Augmented Hateful Meme Detection

Created by
  • Haebom

Author

Jingbiao Mei, Jinghong Chen, Guangyu Yang, Weizhe Lin, Bill Byrne

Outline

This paper focuses on developing a robust system for automatically detecting memes containing hate speech, a serious problem on the Internet. While large-scale multimodal models (LMMs) have shown promising results, they face challenges such as suboptimal performance and limited cross-domain generalization. To address these challenges, we propose a robust adaptive framework that maintains the general vision-language capabilities of LMMs while improving both within-domain accuracy and cross-domain generalization. The proposed method demonstrates robustness against adversarial attacks compared to existing supervised fine-tuning (SFT) models. Experimental results on six meme classification datasets show that it outperforms existing state-of-the-art models and generates higher-quality evidence, thereby enhancing the model's interpretability.

Takeaways, Limitations

Takeaways:
We present a novel adaptive framework that improves within-domain accuracy and cross-domain generalization in hate meme detection using LMMs.
Ensure strong robustness against hostile attacks.
Achieve performance that surpasses current best-in-class performance (SOTA).
Provides improved model interpretability (generating high-quality evidence).
Limitations:
Further research is needed on the generalization performance of the proposed framework.
Generalization performance evaluation for different types of hate memes is needed.
Performance evaluation and continuous monitoring in real environments are required.
👍