Optimizing matching radius for ride-hailing systems with dual-replay-buffer deep reinforcement learning
J. Gao (TU Delft - Transport, Mobility and Logistics)
Rong Cheng (Eindhoven University of Technology)
Yaoxin Wu (Dalian Maritime University)
Honghao Zhao (Student TU Delft)
W. Mai (TU Delft - Traffic Systems Engineering)
O Cats (TU Delft - Transport and Planning)
More Info
expand_more
Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.
Abstract
The matching radius, defined as the maximum pick-up distance within which waiting riders and idle drivers can be matched, is a critical variable in ride-hailing systems. Optimizing the matching radius can significantly enhance system performance, but determining its optimal value is challenging due to the dynamic nature of ride-hailing environments. The matching radius should adapt to spatial and temporal variations, as well as to real-time fluctuations in supply and demand. To address this challenge, this paper proposes a dual-reply-buffer deep reinforcement learning method for dynamic matching radius optimization. By modeling the matching radius optimization problem as a Markov decision process, the method trains a policy network to adaptively adjust the matching radius in response to changing conditions in the ride-hailing system, thereby improving efficiency and service quality. We validate our method using real-world ride-hailing data from Austin, Texas. Experimental results show that the proposed method outperforms baseline approaches, achieving higher matching rates, shorter average pick-up distances, and better driver utilization across different scenarios.