Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Decentralizing Multi-Agent Reinforcement Learning with Temporal Causal Information

Created by
  • Haebom

Author

Jan Corazza, Hadi Partovi Aria, Hyohun Kim, Daniel Neider, Zhe Xu

Outline

While reinforcement learning (RL) algorithms can allow a single agent to find an optimal policy for a specific task, many real-world problems require the collaboration of multiple agents to achieve a common goal. In distributed multi-agent RL (DMARL), agents learn independently and then combine policies at runtime. However, the combination typically requires constraints on the compatibility of local policies to achieve a global task. In this paper, we study how providing high-level symbolic knowledge to agents can help address unique challenges in this setting, such as privacy constraints, communication limitations, and performance issues. Specifically, we extend the formal tool used to verify the compatibility of local policies among team actions, enabling distributed learning with theoretical guarantees to be used in more scenarios. Furthermore, we experimentally demonstrate that symbolic knowledge of the temporal evolution of events in the environment can significantly accelerate the learning process in DMARL.

Takeaways, Limitations

Takeaways:
Providing high-level symbolic knowledge to DMARL agents can address privacy, communication, and performance issues.
Extending the tool for checking local policy compliance enables distributed learning with theoretical guarantees to be applied to more scenarios.
Symbolic knowledge about the temporal evolution of events in the environment can improve DMARL learning speed.
Limitations:
The paper does not explicitly mention Limitations. (This is a summary of the paper, so Limitations can only be determined by directly checking the paper.)
👍