Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Competition and Attraction Improve Model Fusion

Created by
  • Haebom

Author

Jo ao Abrantes, Robert Tjarko Lange, Yujin Tang

Outline

To overcome the limitations of existing model merging techniques, this paper proposes a novel evolutionary algorithm, M2N2, which leverages the concept of natural niches. M2N2 explores diverse model combinations and improves performance through three key features: dynamically adjusting model parameter merging boundaries, maintaining diversity, and selecting promising model pairs. Experimental results show that M2N2 achieves comparable performance to CMA-ES more efficiently by training an MNIST classifier model from scratch, and also demonstrates state-of-the-art performance in merging language and image generation models. Notably, it demonstrates robustness and diversity, preserving important model features that are not explicitly optimized by the fitness function.

Takeaways, Limitations

Takeaways:
We present a novel method for training models from scratch through model merging.
It is more efficient than existing model merging techniques and allows for exploration of diverse model combinations.
It can be applied to various fields such as language and image generation models and achieves state-of-the-art performance.
It also shows robustness in maintaining features that are not explicitly optimized by the fitness function.
Limitations:
Further research may be needed on parameter optimization of the M2N2 algorithm.
Further experiments with different datasets and model architectures are needed.
Due to the nature of evolutionary algorithms, computational costs can be relatively high (although they are claimed to be more efficient than CMA-ES).
👍