Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Learning to Decide with Just Enough: Information-Theoretic Context Summarization for CMDPs

Created by
  • Haebom

Author

Peidong Liu, Junjiang Lin, Shaowen Wang, Yao Xu, Haiqing Li, Xuhao Xie, Siyi Wu, Hao Li

Outline

In the contextual Markov Decision Processes (CMDP) environment, we propose an information-theoretic summarization approach that leverages a large-scale language model (LLM) to compress high-dimensional/unstructured context into a low-dimensional semantic summary. This method augments state by reducing redundancy while preserving crucial decision-making clues. Based on the concept of approximate context sufficiency, we provide a first-of-its-kind regret bound and delay-entropy tradeoff characterization for CMDP. It outperforms existing methods on various benchmarks, improving reward, success rate, and sample efficiency while reducing latency and memory usage.

Takeaways, Limitations

Takeaways:
LLM-based summarization provides a scalable and interpretable solution for efficient decision-making in context-rich, resource-constrained environments.
We provide the first regret bound and delay-entropy tradeoff characterization for CMDP.
Demonstrated performance improvements in a variety of environments.
Limitations:
Limitations, as stated in the paper, is not directly presented (limited to the information provided).
👍