Projects per year
Abstract
Large language models (LLMs) exhibit varying strengths and weaknesses across different tasks, prompting recent studies to explore the benefits of ensembling models to leverage their complementary advantages. However, existing LLM ensembling methods often overlook model compatibility and struggle with inefficient alignment of probabilities across the entire vocabulary. In this study, we empirically investigate the factors influencing ensemble performance, identifying model performance, vocabulary size, and response style as key determinants, revealing that compatibility among models is essential for effective ensembling. This analysis leads to the development of a simple yet effective model selection strategy that identifies compatible models. Additionally, we introduce the UNIon Top-k Ensembling (UNITE), a novel approach that efficiently combines models by focusing on the union of the top-k tokens from each model, thereby avoiding the need for full vocabulary alignment and reducing computational overhead. Extensive evaluations across multiple benchmarks demonstrate that UNITE significantly enhances performance compared to existing methods, offering a more efficient framework for LLM ensembling. The code is available at https://github.com/starrYYxuan/UniTE
| Original language | English |
|---|---|
| Title of host publication | International Conference on Representation Learning 2025 (ICLR 2025) |
| Editors | Y. Yue, A. Garg, N. Peng, F. Sha, R. Yu |
| Publisher | International Conference on Learning Representations, ICLR |
| Pages | 1-19 |
| Number of pages | 19 |
| ISBN (Print) | 9798331320850 |
| Publication status | Published - Jun 2025 |
| Event | 13th International Conference on Learning Representations (ICLR 2025) - Singapore EXPO, Singapore, Singapore Duration: 24 Apr 2025 → 28 Apr 2025 https://iclr.cc/Conferences/2025 |
Publication series
| Name | International Conference on Learning Representations |
|---|
Conference
| Conference | 13th International Conference on Learning Representations (ICLR 2025) |
|---|---|
| Abbreviated title | ICLR 2025 |
| Place | Singapore |
| City | Singapore |
| Period | 24/04/25 → 28/04/25 |
| Internet address |
Bibliographical note
Information for this record is supplemented by the author(s) concerned.Funding
This work was supported in part by the Research Grants Council of the Hong Kong SAR under Grant GRF 11217823 and Collaborative Research Fund C1042-23GF, the National Natural Science Foundation of China under Grant 62371411, InnoHK initiative, the Government of the HKSAR, Laboratory for AI-Powered Financial Technologies.
Research Keywords
- Model ensembling
- LLM
RGC Funding Information
- RGC-funded
Fingerprint
Dive into the research topics of 'DETERMINE-THEN-ENSEMBLE: NECESSITY OF TOP-k UNION FOR LARGE LANGAUGE MADEL ENSEMBLING'. Together they form a unique fingerprint.Projects
- 1 Active
-
GRF: Towards Building An Adaptive Distributed Computation Framework for Massive Context Interplay
SONG, L. (Principal Investigator / Project Coordinator) & LAN, T. (Co-Investigator)
1/01/24 → …
Project: Research