Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Parse Trees Guided LLM Prompt Compression

Created by
  • Haebom

Author

Wenhao Mao, Chengbin Hou, Tianyu Zhang, Xinyu Lin, Ke Tang, Hairong Lv.

Outline

This paper explores how to provide rich context to improve the performance of large-scale language models (LLMs). To address the increased computational cost of long prompts and the limited input size of LLMs, we propose PartPrompt, a novel selective compression method that overcomes the limitations of existing generative and selective compression methods. PartPrompt utilizes a linguistic rule-based syntax tree to compute the information entropy of each node and, based on this, constructs a global tree that considers the hierarchical structure (dependencies among sentences, paragraphs, and sections). It adjusts node values through bottom-up and top-down propagation on the global tree, and then compresses prompts by pruning the tree using a recursive algorithm based on the adjusted node values. Experimental results demonstrate that PartPrompt achieves state-of-the-art performance across diverse datasets, evaluation metrics, compression ratios, and LLMs. It also demonstrates superiority in the cohesiveness of compressed prompts and in extremely long prompt scenarios.

Takeaways, Limitations

Takeaways:
We present a novel prompt compression method that effectively addresses the input limitation and computational cost issues of LLM.
Overcoming the limitations of existing methods by considering language rules and global structures.
Achieving state-of-the-art performance on diverse datasets and LLM.
Improved cohesion of compressed prompts.
Proven effective even for extremely long prompts.
Limitations:
PartPrompt's performance may depend on the accuracy of language rules and syntax analysis.
Further research is needed on generalization performance for specific languages or domains.
There is a need to evaluate the processing efficiency of prompts with very complex structures.
👍