Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

LED-Merging: Mitigating Safety-Utility Conflicts in Model Merging with Location-Election-Disjoint

Created by
  • Haebom

Author

Qianli Ma, Dongrui Liu, Qian Chen, Linfeng Zhang, Jing Shao

Outline

This paper focuses on model merging, a learning-free solution that integrates multiple task-specific models to address the significant computational and data overhead associated with fine-tuning pre-trained large-scale language models (LLMs) for specialized tasks. To address the safety-utility tradeoff (where enhanced generality compromises safety measures) of existing model merging methods, we identify two root causes: neuron misidentification due to simple parameter size-based selection, and neuron interference between tasks during merging. To address these issues, we propose LED-Merging, a three-step framework that identifies task-specific neurons using gradient-based properties, dynamically selects important neurons through multi-model importance fusion, and decouples conflicting updates through parameter isolation. Extensive experiments on Llama-3-8B, Mistral-7B, and Llama2-13B demonstrate that LED-Merging effectively reduces harmful response rates (31.4% reduction on HarmBench for Llama-3-8B-Instruct) while maintaining 95% usability performance (52.39% accuracy on GSM8K). LED-Merging resolves the safety-usefulness tradeoff and provides a lightweight, learning-free paradigm for building robust multi-task LLMs. The code is available on GitHub.

Takeaways, Limitations

Takeaways:
Clearly identify the safety and usability issues of existing model merging methods and propose solutions.
Efficiently build multi-task LLMs without training through LED-Merging.
Achieving two goals simultaneously: reducing adverse response rates and maintaining usability performance.
We present a lightweight, learning-free model merging technique to reduce computational costs.
Limitations:
Further verification of the generalization performance of the proposed method is needed.
Scalability evaluation for various LLM architectures and workloads is needed.
There may be limitations in accurately determining the importance of a neuron for a particular task.
There is room for optimization and improvement in parameter isolation methods.
👍