Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

MuSeD: A Multimodal Spanish Dataset for Sexism Detection in Social Media Videos

Created by
  • Haebom

Author

Laura De Grazia, Pol Pastells, Mauro Vazquez Chas, Desmond Elliott, Danae S anchez Villegas, Mireia Farr us, Mariona Taul e

Outline

This paper presents a multimodal approach for detecting sexism in online video content, particularly on social media platforms like TikTok and Vitut. We introduce a novel Spanish-language multimodal sexism detection dataset, MuSeD (approximately 11 hours of video), and propose an innovative annotation framework that analyzes the contributions of text, speech, and visual modalities. We evaluate various large-scale language models (LLMs) and multimodal LLMs on sexism detection tasks, finding that visual information plays a crucial role in labeling sexist content. While the models effectively detect explicit sexism, they struggle with implicit forms of sexism, such as stereotypes, consistent with low inter-annotator agreement. This underscores the inherent difficulty of identifying implicit sexism, as it relies on social and cultural context.

Takeaways, Limitations

Takeaways:
Introducing MuSeD, a new multimodal dataset for detecting gender discrimination in social media video content.
We present an innovative gender discrimination detection annotation framework that integrates text, voice, and visual modalities.
Evaluating the performance of gender discrimination detection in various LLMs and multimodal LLMs and confirming the importance of visual information.
Emphasizes the difficulty of detecting implicit sexism and the importance of sociocultural context.
Limitations:
The dataset is comprised entirely of Spanish, limiting generalization to other languages.
The model has difficulty detecting implicit sexism (stereotypes, etc.).
There are concerns about the reliability of the annotations, as there are cases where the level of agreement between annotators is low.
👍