DETERMINE-THEN-ENSEMBLE: NECESSITY OF TOP-k UNION FOR LARGE LANGAUGE MADEL ENSEMBLING

Yuxuan Yao, Han Wu*, Mingyang Liu, Sichun Luo, Xiongwei Han, Jie Liu, Zhijiang Guo, Linqi Song*

*Corresponding author for this work

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

Abstract

Large language models (LLMs) exhibit varying strengths and weaknesses across different tasks, prompting recent studies to explore the benefits of ensembling models to leverage their complementary advantages. However, existing LLM ensembling methods often overlook model compatibility and struggle with inefficient alignment of probabilities across the entire vocabulary. In this study, we empirically investigate the factors influencing ensemble performance, identifying model performance, vocabulary size, and response style as key determinants, revealing that compatibility among models is essential for effective ensembling. This analysis leads to the development of a simple yet effective model selection strategy that identifies compatible models. Additionally, we introduce the UNIon Top-k Ensembling (UNITE), a novel approach that efficiently combines models by focusing on the union of the top-k tokens from each model, thereby avoiding the need for full vocabulary alignment and reducing computational overhead. Extensive evaluations across multiple benchmarks demonstrate that UNITE significantly enhances performance compared to existing methods, offering a more efficient framework for LLM ensembling. The code is available at https://github.com/starrYYxuan/UniTE
Original languageEnglish
Title of host publicationInternational Conference on Representation Learning 2025 (ICLR 2025)
EditorsY. Yue, A. Garg, N. Peng, F. Sha, R. Yu
PublisherInternational Conference on Learning Representations, ICLR
Pages1-19
Number of pages19
ISBN (Print)9798331320850
Publication statusPublished - Jun 2025
Event13th International Conference on Learning Representations (ICLR 2025) - Singapore EXPO, Singapore, Singapore
Duration: 24 Apr 202528 Apr 2025
https://iclr.cc/Conferences/2025

Publication series

NameInternational Conference on Learning Representations

Conference

Conference13th International Conference on Learning Representations (ICLR 2025)
Abbreviated titleICLR 2025
PlaceSingapore
CitySingapore
Period24/04/2528/04/25
Internet address

Bibliographical note

Information for this record is supplemented by the author(s) concerned.

Funding

This work was supported in part by the Research Grants Council of the Hong Kong SAR under Grant GRF 11217823 and Collaborative Research Fund C1042-23GF, the National Natural Science Foundation of China under Grant 62371411, InnoHK initiative, the Government of the HKSAR, Laboratory for AI-Powered Financial Technologies.

Research Keywords

  • Model ensembling
  • LLM

RGC Funding Information

  • RGC-funded

Fingerprint

Dive into the research topics of 'DETERMINE-THEN-ENSEMBLE: NECESSITY OF TOP-k UNION FOR LARGE LANGAUGE MADEL ENSEMBLING'. Together they form a unique fingerprint.

Cite this