[공지사항]을 빙자한 안부와 근황 
Show more

Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Learning Nonlinear Causal Reductions to Explain Reinforcement Learning Policies

Created by
  • Haebom

Author

Armin Keki c, Jan Schneider, Dieter B uchler, Bernhard Sch olkopf, Michel Besserve

Outline

Determining the causes of success or failure of reinforcement learning (RL) policies is a challenging problem due to the complex and high-dimensional nature of agent-environment interactions. This study takes a causal perspective to explain the behavior of RL policies by considering states, actions, and rewards as variables in a low-level causal model. We introduce random perturbations to actions during policy execution and observe their effects on the cumulative rewards to learn a simplified high-level causal model that describes these relationships. To this end, we develop a nonlinear causal model reduction framework that guarantees approximate intervention consistency, meaning that the simplified high-level model responds to interventions in a similar way to the original complex system. We demonstrate that there exists a unique solution that achieves exact intervention consistency for a class of nonlinear causal models, ensuring that the learned explanations reflect meaningful causal patterns. Experiments on synthetic causal models and real-world RL tasks, including pendulum control and robotic table tennis, demonstrate that the proposed approach can uncover important behavioral patterns, biases, and failure modes in trained RL policies.

Takeaways, Limitations

Takeaways:
We present a novel framework that provides causal explanations for the behavior of RL policies.
Simplify complex systems and increase interpretability through nonlinear causal model reduction.
Increases the reliability of learned explanations by ensuring approximate intervention consistency.
We validate its effectiveness on synthetic data and real-world RL tasks.
Limitations:
The applicability of the proposed framework may be limited to certain kinds of nonlinear causal models.
Computational cost in high-dimensional state spaces can be high.
Further research is needed on generalization performance in real-world RL tasks.
👍