Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

LLMs Can't Handle Peer Pressure: Crumbling under Multi-Agent Social Interactions

Created by
  • Haebom

Author

Maojia Song, Tej Deep Pala, Weisheng Jin, Amir Zadeh, Chuan Li, Dorien Herremans, Souzanya Poria

Outline

This paper analyzes the trust formation, misinformation resistance, and peer input integration capabilities of a large-scale language model (LLM), which is used as a component of collaborative intelligence in multi-agent systems (MAS). Unlike previous studies that primarily focus on groupthink, this paper delves into these aspects, which are crucial for achieving collective intelligence under complex social dynamics. To achieve this, we present KAIROS, a benchmark that simulates a quiz competition with peer agents of varying trust levels, allowing for fine-tuning various conditions such as expert-novice roles, noisy crowds, and adversarial peers. The LLM systematically investigates the impact of trust, peer behavior, and self-confidence on decision-making by collecting both past interactions and current peer responses. We evaluate prompting, supervised learning fine-tuning, and reinforcement learning (GRPO) as mitigation strategies across multiple models. Our results show that GRPO, which combines a multi-agent context, outcome-based rewards, and unconstrained inference, achieves the best performance but is less robust to social influence than the baseline model. The code and dataset are publicly available.

Takeaways, Limitations

Takeaways:
We provide an in-depth understanding of LLM's trust formation, misinformation resistance, and peer input aggregation mechanisms in multi-agent systems.
We present a new benchmark, KAIROS, which can control various conditions.
We evaluate the effectiveness of various mitigation strategies, including GRPO, and propose the optimal strategy.
We present directions for improving the performance of LLM-based multi-agent systems.
Limitations:
GRPO achieves the best performance but has reduced robustness to social influences.
Further research is needed to determine the generalizability of the KAIROS benchmark.
Further research is needed to determine its applicability to real-world scenarios.
👍