REFUSION: IMPROVING NATURAL LANGUAGE UNDERSTANDING WITH COMPUTATION-EFFICIENT RETRIEVAL REPRESENTATION FUSION

Shangyu Wu, Ying Xiong, Yufei Cui*, Xue Liu, Buzhou Tang, Tei-Wei Kuo, Chun Jason Xue

*Corresponding author for this work

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

1 Citation (Scopus)

Abstract

Retrieval-based augmentations (RA) incorporating knowledge from an external database into language models have greatly succeeded in various knowledge-intensive (KI) tasks. However, integrating retrievals in non-knowledge-intensive (NKI) tasks is still challenging. Existing works focus on concatenating retrievals with inputs to improve model performance. Unfortunately, the use of retrieval concatenation-based augmentations causes an increase in the input length, substantially raising the computational demands of attention mechanisms. This paper proposes a new paradigm of RA named ReFusion, a computation-efficient Retrieval representation Fusion with bi-level optimization. Unlike previous works, ReFusion directly fuses the retrieval representations into the hidden states of models. Specifically, ReFusion leverages an adaptive retrieval integrator to seek the optimal combination of the proposed ranking schemes across different model layers. Experimental results demonstrate that the proposed ReFusion can achieve superior and robust performance in various NKI tasks. © 2024 12th International Conference on Learning Representations, ICLR 2024. All rights reserved.
Original languageEnglish
Title of host publicationThe Twelfth International Conference on Learning Representations, ICLR 2024
PublisherInternational Conference on Learning Representations, ICLR
Number of pages16
Publication statusPublished - 2024
Event12th International Conference on Learning Representations (ICLR 2024) - Messe Wien Exhibition and Congress Center, Vienna, Austria
Duration: 7 May 202411 May 2024
https://iclr.cc/Conferences/2024
https://openreview.net/group?id=ICLR.cc/2024/Conference

Publication series

NameInternational Conference on Learning Representations, ICLR

Conference

Conference12th International Conference on Learning Representations (ICLR 2024)
Country/TerritoryAustria
CityVienna
Period7/05/2411/05/24
Internet address

Funding

The work described in this paper was supported by a grant from the Research Grants Council of the Hong Kong Special Administrative Region, China (Project No. CityU 11209122).

Fingerprint

Dive into the research topics of 'REFUSION: IMPROVING NATURAL LANGUAGE UNDERSTANDING WITH COMPUTATION-EFFICIENT RETRIEVAL REPRESENTATION FUSION'. Together they form a unique fingerprint.

Cite this