RecRanker: Instruction Tuning Large Language Model as Ranker for Top-k Recommendation

SICHUN LUO, BOWEI HE, HAOHAN ZHOU, WEI SHAO, YANLIN QI, YINYA HUANG, AOJUN ZHOU, YUXUAN YAO, ZONGPENG LI, YUANZHANG XIAO, MINGJIE ZHAN, LINQI SONG*

*Corresponding author for this work

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

Abstract

Large Language Models (LLMs) have demonstrated remarkable capabilities and have been extensively deployed across various domains, including recommender systems. Prior research has employed specialized prompts to leverage the in-context learning capabilities of LLMs for recommendation purposes. More recent studies have utilized instruction tuning techniques to align LLMs with human preferences, promising more effective recommendations. However, existing methods suffer from several limitations. The full potential of LLMs is not fully elicited due to low-quality tuning data and the overlooked integration of conventional recommender signals. Furthermore, LLMs may generate inconsistent responses for different ranking tasks in the recommendation, potentially leading to unreliable results.

In this paper, we introduce RecRanker, tailored for instruction tuning LLMs to serve as the Ranker for top-k Recommendations. Specifically, we introduce importance-aware sampling, clustering-based sampling, and penalty for repetitive sampling for sampling high-quality, representative, and diverse training data. To enhance the prompt, we introduce a position shifting strategy to mitigate position bias and augment the prompt with auxiliary information from conventional recommendation models, thereby enriching the contextual understanding of the LLM. Subsequently, we utilize the sampled data to assemble an instruction-tuning dataset with the augmented prompts comprising three distinct ranking tasks: pointwise, pairwise, and listwise rankings. We further propose a hybrid ranking method to enhance the model performance by ensembling these ranking tasks. Our empirical evaluations demonstrate the effectiveness of our proposed RecRanker in both direct and sequential recommendation scenarios. © 2024 Copyright held by the owner/author(s).
Original languageEnglish
Article number113
JournalACM Transactions on Information Systems
Volume43
Issue number5
Online published29 Nov 2024
DOIs
Publication statusPublished - 10 Jul 2025

Bibliographical note

Full text of this publication does not contain sufficient affiliation information. With consent from the author(s) concerned, the Research Unit(s) information for this record is based on the existing academic department affiliation of the author(s)

Funding

This work was supported in part by the Research Grants Council of the Hong Kong SAR under Grant GRF 11217823 and Collaborative Research Fund C1042-23GF, the National Natural Science Foundation of China under Grant 62371411, InnoHK initiative, the Government of the HKSAR, Laboratory for AI-Powered Financial Technologies.

Research Keywords

  • Recommender System
  • Large Language Model
  • Instruction Tuning
  • Ranking

Fingerprint

Dive into the research topics of 'RecRanker: Instruction Tuning Large Language Model as Ranker for Top-k Recommendation'. Together they form a unique fingerprint.

Cite this