[공지사항]을 빙자한 안부와 근황 
Show more

Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Judging with Many Minds: Do More Perspectives Mean Less Prejudice? On Bias Amplifications and Resistance in Multi-Agent Based LLM-as-Judge

Created by
  • Haebom

Author

Chiyu Ma, Enpei Zhang, Yilun Zhao, Wenjun Liu, Yaning Jia, Peijun Qing, Lin Shi, Arman Cohan, Yujun Yan, Soroush Vosoughi

Outline

This paper systematically analyzes the impact of inherent bias in a multi-agent extension of the LLM-as-Judge approach (multi-agent argumentation and meta-evaluation) that uses large-scale language models (LLMs) as evaluators. By evaluating four types of bias (position bias, detail bias, thought process bias, and consensus bias) in both the multi-agent argumentation and LLM-as-Meta-Judge frameworks, we find that the argumentation framework significantly amplifies and persists bias after the initial argumentation, while the meta-evaluation approach is more resistant to bias. In addition, we show that adding an unbiased agent using PINE, a single-agent bias reduction method, is effective in reducing bias in the argumentation setting, but less effective in the meta-evaluation setting. In conclusion, this study comprehensively studies the behavior of bias in the multi-agent LLM-as-Judge system and highlights the need for targeted bias mitigation strategies in collaborative evaluation settings.

Takeaways, Limitations

Takeaways:
Provides an in-depth understanding of how different types of bias manifest themselves in multi-agent LLM-as-Judge systems.
We reveal differences in the bias resistance of multi-agent argument frameworks and meta-evaluation frameworks.
We analyze the effectiveness of applying single-agent bias reduction techniques to multi-agent systems and show the difference in effectiveness depending on the settings.
It highlights the need to develop effective bias mitigation strategies in collaborative evaluation settings.
Limitations:
The types of biases analyzed may be limited. Additional research is needed on other types of bias.
Since the results are for a specific LLM and dataset, further validation of generalizability is needed.
There is a lack of application and comparative analysis of other bias reduction techniques besides PINE.
The complexity of multi-agent systems may require a more in-depth analysis of the causes and mechanisms of bias.
👍