Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Can an Individual Manipulate the Collective Decisions of Multi-Agents?

Created by
  • Haebom

Author

Fengyuan Liu, Rui Zhao, Shuo Chen, Guohao Li, Philip Torr, Lei Han, Jindong Gu

Outline

Individual large-scale language models (LLMs) have demonstrated outstanding performance in various fields, and collaboratively coordinated multi-agent systems have enhanced decision-making and reasoning capabilities. We raise the question of whether an attacker can generate adversarial samples capable of misleading the collective decisions of a multi-agent system, even when only one agent is aware of the information. We propose a framework, called M-Spoiler, which formalizes this approach as an incomplete information game. It simulates agent interactions to generate adversarial samples and manipulates the collaborative decision-making process of the target system. M-Spoiler introduces a robust agent that simulates potential robust responses from agents in the target system, thereby assisting in the optimization of adversarial samples. Extensive experiments demonstrate the risks posed by single-agent knowledge to multi-agent systems and demonstrate the effectiveness of the proposed attack framework. Furthermore, we explore defense mechanisms and demonstrate that the proposed attack framework is more robust than existing methods.

Takeaways, Limitations

Takeaways:
We show that in a multi-agent system, knowledge of even a single agent can influence the collective decisions of the system.
The M-Spoiler framework presents an effective methodology for attacking multi-agent systems by generating adversarial samples.
The fact that M-Spoiler's attacks remain powerful despite various defense mechanisms highlights the need to develop new defense strategies.
Limitations:
Focusing on specific attack scenarios may limit exploration of other attack methodologies.
Results may vary depending on the specific tasks and system settings used in the experiment.
Exploration of defense mechanisms is in its infancy and further research is needed.
👍