Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Memorization in Fine-Tuned Large Language Models

Created by
  • Haebom

Author

Danil Savine

Outline

This study investigates the memory mechanisms and factors in fine-tuned large-scale language models (LLMs), focusing on privacy concerns in the healthcare field. Using the PHEE dataset of pharmacovigilance events, we examine how various aspects of the fine-tuning process affect the model's tendency to memorize training data. We detect memorized data using two main methods: membership inference attacks and a generation task using prompt prefixes. We analyze the application of different weight matrices in transformer architectures, the relationship between perplexity and memorization, and the effect of increasing rank in Low-Rank Adaptation (LoRA) fine-tuning. Key findings include: (1) the value and output matrices contribute more to memorization than the query and key matrices; (2) lower perplexity in fine-tuned models leads to increased memorization; and (3) higher LoRA ranks lead to increased memorization, but with diminishing returns. These findings provide insight into the trade-off between model performance and privacy risks in fine-tuned LLMs. The findings of this study provide Takeaways guidance for developing more effective and responsible strategies for applying large-scale language models while managing data privacy concerns.

Takeaways, Limitations

Takeaways:
By revealing that specific weight matrices (values, output matrices) in fine-tuned LLM contribute more to memorizing training data, we contribute to establishing a privacy-conscious fine-tuning strategy.
Helps balance model performance and privacy by identifying the correlation between embarrassment and memorization.
Contributes to the development of an efficient LoRA parameter adjustment strategy by presenting the change in memorization patterns according to the increase in LoRA class.
Focusing on privacy issues in the healthcare field, provides important Takeaways for practical applications.
Limitations:
Limitations on generalizability using only one PHEE dataset.
Further research is needed to determine whether this generalization applies to various LLM architectures or fine-tuning techniques.
Since we evaluated memorization using only two methods, membership inference attacks and generation tasks, other memorization measures should be considered.
Rather than presenting specific privacy strategies, the focus is on increasing understanding of memorization mechanisms.
👍