Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Explain Less, Understand More: Jargon Detection via Personalized Parameter-Efficient Fine-tuning

Created by
  • Haebom

Author

Bohao Wu, Qingyun Wang, Yue Guo

Outline

This paper presents a method for detecting specialized terms and personalizing explanations to enable readers with diverse backgrounds to understand specialized documents. Since existing user-specific fine-tuning approaches require significant annotation effort and computational resources, this paper explores efficient and scalable personalization strategies. Specifically, we explore two strategies: lightweight fine-tuning using Low-Rank Adaptation (LoRA) on open-source models and personalized prompting, which adjusts model behavior at inference time. We also study a hybrid approach that combines limited annotation data with user background signals from unsupervised learning. Experimental results show that the personalized LoRA model outperforms GPT-4 by 21.4% in F1 score and the best-performing oracle baseline model by 8.3%. Furthermore, it achieves similar performance using only 10% of the annotated training data, demonstrating its practicality even in resource-constrained environments.

Takeaways, Limitations

Takeaways:
This is the first study to systematically investigate an efficient and resource-efficient term detection personalization method using open-source language models.
We achieved performance that surpassed GPT-4 through LoRA-based lightweight fine-tuning.
It contributes to building practical, scalable, user-adaptive NLP systems by maintaining high performance even with limited data.
Limitations:
Further validation of the generalizability of the dataset used in the study is needed.
Further performance evaluations are needed for different types of terminology and documents.
Research is needed to further improve the effectiveness of personalized prompting strategies.
👍