[공지사항]을 빙자한 안부와 근황 
Show more

Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Dynamic Context Tuning for Retrieval-Augmented Generation: Enhancing Multi-Turn Planning and Tool Adaptation

Created by
  • Haebom

Author

Jubin Abhishek Soni, Amit Anand, Rajesh Kumar Pandey, Aniket Abhishek Soni

Outline

Existing large-scale language models (LLMs) based on Retrieval-Augmented Generation (RAG) are limited to static single-turn interactions and a fixed set of tools, making them unsuitable for dynamic environments such as healthcare and smart homes. In this paper, we present Dynamic Context Tuning (DCT), a lightweight framework that extends RAG to support multi-turn conversations and changing tool environments without retraining. DCT integrates an attention-based context cache to track relevant historical information, LoRA-based retrieval to dynamically select domain-specific tools, and efficient context compression to keep inputs within LLM context constraints. Synthetic and real-world benchmark experiments show that DCT improves planning accuracy by 14% and reduces hallucinations by 37%, while achieving GPT-4 performance at a much lower cost. Furthermore, DCT generalizes to previously unseen tools, enabling scalable and adaptable AI assistants in diverse dynamic environments.

Takeaways, Limitations

Takeaways:
We present DCT, a lightweight framework for RAG systems supporting multi-turn conversations and changing tool environments.
Performance improvement through improved planning accuracy and reduced hallucinations (14% planning accuracy, 37% reduced hallucinations).
Achieving GPT-4 level performance at a much lower cost.
Generalizable to previously unseen tools, suggesting the possibility of implementing scalable and adaptable AI assistants in diverse and dynamic environments.
Limitations:
Further research is needed to determine the generalizability of the experimental results presented in this paper.
Additional application and performance evaluation for various real environments is needed.
Additional verification of the scalability and stability of DCT is needed.
👍