Dynamic Resource Routing using Real-Time Dynamic Programming

Dynamic Resource Routing using Real-Time Dynamic Programming

Sebastian Schmoll, Matthias Schubert

Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence
Main track. Pages 4822-4828. https://doi.org/10.24963/ijcai.2018/670

Acquiring available resources in stochastic environments becomes more and more important to future mobility. For instance, cities like Melbourne, Canberra and San Francisco install sensors that detect in real-time whether a parking spot (resource) is available or not. In such environments, the current state of the resources may be fully observable, although the future development is stochastic. In order to reduce the traffic, such cities want to fully exploit parking spots, such that the amount of searching cars is minimized. Thus, we formulate a problem setting where the expected seek time for each driver is minimized. This problem can be modeled by a Markov Decision Process (MDP) and solved using standard algorithms. In this paper, we focus on the setting, where pre-computation is not possible and search policies have to be computed on the fly. Our approach is based on state-of-the-art Real-Time Dynamic Programming (RTDP) approaches. However, standard RTDP approaches do not perform well on this specific problem setting as shown in our experiments. We introduce adapted bounds and approximations that exploit the specific nature of the problem in order to improve the performance significantly.
Keywords:
Planning and Scheduling: Markov Decisions Processes
Planning and Scheduling: Planning under Uncertainty
Planning and Scheduling: Real-time Planning
Planning and Scheduling: Applications of Planning