[공지사항]을 빙자한 안부와 근황 
Show more

Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Prompt Perturbations Reveal Human-Like Biases in LLM Survey Responses

Created by
  • Haebom

Author

Jens Rupprecht, Georg Ahnert, Markus Strohmaier

Outline

This paper investigates the reliability and vulnerability to response bias of large-scale language models (LLMs) as surrogates for human subjects in social science surveys. Using the World Values Survey (WVS) questionnaire, we conducted over 167,000 mock interviews with nine different LLMs, applying 11 changes to the question format and response option structure. We find that LLMs are not only vulnerable to change, but also exhibit consistent recency bias across all models, with varying strengths, and over-prefer the last response option presented. Although larger models are generally more robust, all models are still sensitive to semantic changes such as rephrasing and complex changes. By applying a series of changes, we find that LLMs partially match the survey response biases observed in humans. This highlights the importance of prompt design and robustness testing when generating synthetic survey data using LLMs.

Takeaways, Limitations

Takeaways:
Shows that LLM is vulnerable to response bias when applied to social science surveys.
Confirming the recency bias that consistently appears in LLMs.
Emphasize the importance of prompt design and robustness testing when generating synthetic survey data using LLM.
The larger the LLM size, the higher the robustness generally.
We confirm that LLM's response bias partially matches human response bias.
Limitations:
The type of LLM used in this study and the characteristics of the WVS questions may affect the generalizability of the study results.
Further research is needed on more diverse types of survey questions and perturbations.
Further research is needed to determine how to completely eliminate response bias in LLMs.
👍