Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Offline Fictitious Self-Play for Competitive Games

Created by
  • Haebom

Author

Jingxiao Chen, Weiji Xie, Weinan Zhang, Yong yu, Ying Wen

Outline

OFF-FSP is an offline reinforcement learning algorithm that enables policy improvement using only a fixed dataset, developed specifically for competitive game environments. This algorithm simulates various opponents through virtual interactions in situations where the game structure is unknown and utilizes an offline self-play learning framework. Furthermore, to overcome incomplete data coverage, it approximates a Nash equilibrium by combining single-agent offline reinforcement learning with fictional self-play. Experiments on matrix games, poker, board games, and real-world human-robot competition tasks demonstrate that OFF-FSP outperforms existing methods.

Takeaways, Limitations

A Practical Model-Free Offline Reinforcement Learning Algorithm for Competitive Games
Approximating Nash Equilibrium by Combining Offline Self-Play Learning Framework and Fictional Self-Play
Demonstrated superior performance in a variety of environments, including matrix games, poker, board games, and real-world human-robot competition tasks.
Contributes to solving the problem of incomplete coverage of offline datasets.
The performance of an algorithm depends heavily on the quality and diversity of the dataset.
The complexity of the algorithm may result in high computational costs.
When applied in real environments, additional efforts in data collection and preprocessing may be required.
👍