NFL : Robust Learned Index via Distribution Transformation
Research output: Journal Publications and Reviews (RGC: 21, 22, 62) › 21_Publication in refereed journal › peer-review
Author(s)
Related Research Unit(s)
Detail(s)
Original language | English |
---|---|
Pages (from-to) | 2188–2200 |
Journal / Publication | Proceedings of the VLDB Endowment |
Volume | 15 |
Issue number | 10 |
Publication status | Published - Jun 2022 |
Conference
Title | 48th International Conference on Very Large Data Bases, VLDB 2022 |
---|---|
Place | Australia |
City | Sydney |
Period | 5 - 9 September 2022 |
Link(s)
Abstract
Recent works on learned index open a new direction for the indexing field. The key insight of the learned index is to approximate
the mapping between keys and positions with piece-wise linear
functions. Such methods require partitioning key space for a better
approximation. Although lots of heuristics are proposed to improve
the approximation quality, the bottleneck is that the segmentation
overheads could hinder the overall performance.
This paper tackles the approximation problem by applying a distribution transformation to the keys before constructing the learned index. A two-stage Normalizing-Flow-based Learned index framework (NFL) is proposed, which first transforms the original complex key distribution into a near-uniform distribution, then builds a learned index leveraging the transformed keys. For effective distribution transformation, we propose a Numerical Normalizing Flow (Numerical NF). Based on the characteristics of the transformed keys, we propose a robust After-Flow Learned Index (AFLI). To validate the performance, comprehensive evaluations are conducted on both synthetic and real-world workloads, which shows that the proposed NFL produces the highest throughput and the lowest tail latency compared to the state-of-the-art learned indexes.
This paper tackles the approximation problem by applying a distribution transformation to the keys before constructing the learned index. A two-stage Normalizing-Flow-based Learned index framework (NFL) is proposed, which first transforms the original complex key distribution into a near-uniform distribution, then builds a learned index leveraging the transformed keys. For effective distribution transformation, we propose a Numerical Normalizing Flow (Numerical NF). Based on the characteristics of the transformed keys, we propose a robust After-Flow Learned Index (AFLI). To validate the performance, comprehensive evaluations are conducted on both synthetic and real-world workloads, which shows that the proposed NFL produces the highest throughput and the lowest tail latency compared to the state-of-the-art learned indexes.
Bibliographic Note
Research Unit(s) information for this publication is provided by the author(s) concerned.
Citation Format(s)
NFL : Robust Learned Index via Distribution Transformation. / Wu, Shangyu; Cui, Yufei; Yu, Jinghuan et al.
In: Proceedings of the VLDB Endowment, Vol. 15, No. 10, 06.2022, p. 2188–2200.Research output: Journal Publications and Reviews (RGC: 21, 22, 62) › 21_Publication in refereed journal › peer-review