Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Improved Monte Carlo Planning via Causal Disentanglement for Structurally-Decomposed Markov Decision Processes

Created by
  • Haebom

Author

Larkin Liu, Shiqi Liu, Yinruo Hua, Matej Jusup

Outline

To overcome the limitations of Markov Decision Processes (MDPs), we propose Structurally Decomposed MDP (SD-MDP) utilizing causal structures. SD-MDP partitions the temporal causal graph of an MDP into independent components, achieving dimensionality reduction and computational efficiency. For resource allocation problems, it reduces the optimization problem to a fractional knapsack problem with log-linear complexity of O(T log T) for the time horizon T. This reduces the complexity to O(T log T), which is superior to conventional probabilistic programming methods with polynomial complexity, and it is valid in high-dimensional spaces regardless of the size of the state-action space. Furthermore, SD-MDP integrates with Monte Carlo Tree Search (MCTS) to achieve higher expected rewards under limited simulation budgets and eliminates the simple regret boundary. It demonstrates superior policy performance compared to benchmarks in logistics and finance.

Takeaways, Limitations

Takeaways:
We improved computational efficiency by leveraging the causal structure of MDP.
It achieves log-linear complexity in resource allocation problems, outperforming existing methods.
It can also be applied in high-dimensional spaces.
Integration with MCTS has shown performance improvements.
Positive results were achieved in the logistics and finance sectors.
Limitations:
The specific Limitations was not specified in the paper. (Only the abstract of the paper was provided, so the specific details of Limitations are not available.)
👍