Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Implicit Safe Set Algorithm for Provably Safe Reinforcement Learning

Created by
  • Haebom

Author

Weiye Zhao, Feihan Li, Changliu Liu

Outline

This paper presents the implicit safety set algorithm, a model-free safety control algorithm that guarantees the safety of deep reinforcement learning (DRL) agents. While existing DRL methods struggle to guarantee safety, our algorithm generates safety indices (barrier certificates) and safety control laws using only black-box dynamic functions (e.g., digital twin simulators). We theoretically prove convergence to a safe set within finite time and forward invariance for both continuous and discrete-time systems. Furthermore, we demonstrate our performance on the Safety Gym benchmark, achieving a cumulative reward of 95% ± 9%, outperforming the existing state-of-the-art safe DRL method, without safety violations. Furthermore, we demonstrate scalability to high-dimensional systems via parallel computing.

Takeaways, Limitations

Takeaways:
A new algorithm is presented to ensure the safety of DRL agents in a model-free manner.
Theoretically prove safety and convergence
Excellent performance verified in the Safety Gym benchmark
Verifying scalability to high-dimensional systems
Limitations:
Further validation of the algorithm's real-world application is needed.
Generalizability studies are needed for various safety constraints.
Further analysis is needed on the computational complexity and efficiency of implicit safety set algorithms.
👍