This paper investigates how a Vision-Language Model (VLM) reflects gender stereotypes through embeddings between facial images and phrases describing occupations and activities. Using a dataset of 220 gender-differentiated facial images and 150 phrases (in six categories: emotional labor, cognitive labor, domestic labor, technical labor, professional labor, and manual labor), we compute a gender relevance score by calculating the difference in cosine similarity between phrases for male and female image embeddings. We present a robust framework for assessing gender bias by calculating confidence intervals using bootstrapping and comparing these results to the expected value in the absence of gender structure. Consequently, we provide gender relevance maps for phrases and categories.