Fed-LTD : Towards Cross-Platform Ride Hailing via Federated Learning to Dispatch

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

28 Scopus Citations
View graph of relations

Author(s)

  • Yansheng Wang
  • Yongxin Tong
  • Ziyao Ren
  • Yi Xu
  • Guobin Wu
  • Weifeng Lv

Detail(s)

Original languageEnglish
Title of host publicationKDD '22: Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining
PublisherAssociation for Computing Machinery
Pages4079-4089
ISBN (print)9781450393850
Publication statusPublished - Aug 2022
Externally publishedYes

Publication series

NameProceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining

Conference

Title28th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD 2022)
LocationWashington DC Convention Center
PlaceUnited States
CityWashington DC
Period14 - 18 August 2022

Abstract

Learning based order dispatching has witnessed tremendous success in ride hailing. However, the success halts within individual ride hailing platforms because sharing raw order dispatching data across platforms may leak user privacy and business secrets. Such data isolation not only impairs user experience but also decreases the potential revenues of the platforms. In this paper, we advocate federated order dispatching for cross-platform ride hailing, where multiple platforms collaboratively make dispatching decisions without sharing their local data. Realizing this concept calls for new federated learning strategies that tackle the unique challenges on effectiveness, privacy and efficiency in the context of order dispatching. In response, we devise Federated Learning-to-Dispatch (Fed-LTD), a framework that allows effective order dispatching by sharing both dispatching models and decisions while providing privacy protection of raw data and high efficiency. We validate Fed-LTD via large-scale trace-driven experiments with Didi GAIA dataset. Extensive evaluations show that Fed-LTD outperforms single-platform order dispatching by 10.24% to 54.07% in terms of total revenue. © 2022 ACM.

Research Area(s)

  • federated learning, order dispatching, ride hailing

Citation Format(s)

Fed-LTD: Towards Cross-Platform Ride Hailing via Federated Learning to Dispatch. / Wang, Yansheng; Tong, Yongxin; Zhou, Zimu et al.
KDD '22: Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. Association for Computing Machinery, 2022. p. 4079-4089 (Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining).

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review