Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Style-Preserving Policy Optimization for Game Agents

Created by
  • Haebom

Author

Lingfeng Li, Yunlong Lu, Yongyi Wang, Wenxin Li

Outline

This paper proposes a solution to the problem that reinforcement learning-based game AI focuses on improving skill, while evolutionary algorithm-based methods generate diverse play styles but suffer from poor performance. We present Mixed Proximal Policy Optimization (MPPO), a method that improves the skill of existing low-performing agents while maintaining their unique styles. MPPO integrates loss objectives for online and offline samples and introduces implicit constraints that approximate the demo agent's policy by adjusting the empirical distribution of the samples. Experimental results on environments of various scales demonstrate that MPPO achieves skill levels similar to or better than purely online algorithms while preserving the play styles of the demo agent. Consequently, we present an effective method for generating highly skilled and diverse game agents that contribute to more immersive gaming experiences.

Takeaways, Limitations

Takeaways:
Presenting a method to effectively resolve the trade-off between proficiency and diversity, a limitation of existing reinforcement learning-based game AI.
Demonstrating the feasibility of generating high-performance and diverse playstyle game agents through MPPO.
Contributes to improving the quality of the gaming experience and increasing replay value.
Limitations:
Further research is needed on the generalization performance of the proposed method.
Further validation of MPPO's applicability and efficiency across diverse gaming environments and complexities is needed.
A more in-depth analysis is needed on how the selection and quality of demo agents impact the final outcome.
👍