Abstract
While most current approaches rely on further training techniques, such as fine-tuning or reinforcement learning, to enhance model capacities, model merging stands out for its ability of improving models without requiring any additional training. In this paper, we propose a unified framework for model merging based on low-rank estimation of task vectors without the need for access to the base model, named LORE-MERGING. Our approach is motivated by the observation that task vectors from finetuned models frequently exhibit a limited number of dominant singular values, making low-rank estimations less prone to interference. We implement the method by formulating the merging problem as an optimization problem. Extensive empirical experiments demonstrate the effectiveness of our framework in mitigating interference and preserving task-specific information, thereby advancing the state-of-the-art performance in model merging techniques. © 2025 Association for Computational Linguistics.
| Original language | English |
|---|---|
| Title of host publication | Findings of the Association for Computational Linguistics |
| Subtitle of host publication | EMNLP 2025 |
| Editors | Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng |
| Publisher | Association for Computational Linguistics |
| Pages | 21919-21926 |
| Number of pages | 8 |
| ISBN (Print) | 9798891763357 |
| DOIs | |
| Publication status | Published - Nov 2025 |
| Event | 30th Conference on Empirical Methods in Natural Language Processing (EMNLP 2025) - Suzhou, China Duration: 4 Nov 2025 → 9 Nov 2025 https://aclanthology.org/volumes/2025.emnlp-main/ |
Publication series
| Name | EMNLP - Conference on Empirical Methods in Natural Language Processing, Findings of EMNLP |
|---|
Conference
| Conference | 30th Conference on Empirical Methods in Natural Language Processing (EMNLP 2025) |
|---|---|
| Abbreviated title | 30th EMNLP |
| Place | China |
| City | Suzhou |
| Period | 4/11/25 → 9/11/25 |
| Internet address |
Publisher's Copyright Statement
- This full text is made available under CC-BY 4.0. https://creativecommons.org/licenses/by/4.0/
Fingerprint
Dive into the research topics of 'LORE-MERGING: Exploring Low-Rank Estimation For Large Language Model Merging'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver