STRec : Sparse Transformer for Sequential Recommendations
Research output: Chapters, Conference Papers, Creative and Literary Works › RGC 32 - Refereed conference paper (with host publication) › peer-review
Author(s)
Related Research Unit(s)
Detail(s)
Original language | English |
---|---|
Title of host publication | RecSys '23 |
Subtitle of host publication | Proceedings of the 17th ACM Conference on Recommender Systems |
Editors | Jie Zhang, Li Chen, Shlomo Berkovsky, Min Zhang, Tommaso di Noia, Justin Basilico, Luiz Pizzato, Yang Song |
Publisher | Association for Computing Machinery |
Pages | 101-111 |
ISBN (print) | 979-8-4007-0241-9 |
Publication status | Published - Sept 2023 |
Conference
Title | 17th ACM Conference on Recommender Systems (RecSys 2023) |
---|---|
Location | |
Place | Singapore |
City | |
Period | 18 - 22 September 2023 |
Link(s)
Abstract
With the rapid evolution of transformer architectures, researchers
are exploring their application in sequential recommender systems
(SRSs) and presenting promising performance on SRS tasks compared with former SRS models. However, most existing transformer-based SRS frameworks retain the vanilla attention mechanism,
which calculates the attention scores between all item-item pairs.
With this setting, redundant item interactions can harm the model
performance and consume much computation time and memory.
In this paper, we identify the sparse attention phenomenon in
transformer-based SRS models and propose Sparse Transformer for
sequential Recommendation tasks (STRec) to achieve the efficient
computation and improved performance. Specifically, we replace
self-attention with cross-attention, making the model concentrate
on the most relevant item interactions. To determine these necessary interactions, we design a novel sampling strategy to detect
relevant items based on temporal information. Extensive experimental results validate the effectiveness of STRec, which achieves
the state-of-the-art accuracy while reducing 54% inference time and
70% memory cost. We also provide massive extended experiments
to further investigate the property of our framework.
© 2023 Copyright held by the owner/author(s). Publication rights licensed to ACM.
© 2023 Copyright held by the owner/author(s). Publication rights licensed to ACM.
Citation Format(s)
STRec: Sparse Transformer for Sequential Recommendations. / Li, Chengxi; Wang, Yejing; Liu, Qidong et al.
RecSys '23: Proceedings of the 17th ACM Conference on Recommender Systems. ed. / Jie Zhang; Li Chen; Shlomo Berkovsky; Min Zhang; Tommaso di Noia; Justin Basilico; Luiz Pizzato; Yang Song. Association for Computing Machinery, 2023. p. 101-111.
RecSys '23: Proceedings of the 17th ACM Conference on Recommender Systems. ed. / Jie Zhang; Li Chen; Shlomo Berkovsky; Min Zhang; Tommaso di Noia; Justin Basilico; Luiz Pizzato; Yang Song. Association for Computing Machinery, 2023. p. 101-111.
Research output: Chapters, Conference Papers, Creative and Literary Works › RGC 32 - Refereed conference paper (with host publication) › peer-review