Open Access
2013 Tuning approximate dynamic programming policies for ambulance redeployment via direct search
Matthew S. Maxwell, Shane G. Henderson, Huseyin Topaloglu
Stoch. Syst. 3(2): 322-361 (2013). DOI: 10.1214/10-SSY020

Abstract

In this paper we consider approximate dynamic programming methods for ambulance redeployment. We first demonstrate through simple examples how typical value function fitting techniques, such as approximate policy iteration and linear programming, may not be able to locate a high-quality policy even when the value function approximation architecture is rich enough to provide the optimal policy. To make up for this potential shortcoming, we show how to use direct search methods to tune the parameters in a value function approximation architecture so as to obtain high-quality policies. Direct search is computationally intensive. We therefore use a post-decision state dynamic programming formulation of ambulance redeployment that, together with direct search, requires far less computation with no noticeable performance loss. We provide further theoretical support for the post-decision state formulation of the ambulance-deployment problem by showing that this formulation can be obtained through a limiting argument on the original dynamic programming formulation.

Citation

Download Citation

Matthew S. Maxwell. Shane G. Henderson. Huseyin Topaloglu. "Tuning approximate dynamic programming policies for ambulance redeployment via direct search." Stoch. Syst. 3 (2) 322 - 361, 2013. https://doi.org/10.1214/10-SSY020

Information

Published: 2013
First available in Project Euclid: 11 February 2014

zbMATH: 1295.90099
MathSciNet: MR3353206
Digital Object Identifier: 10.1214/10-SSY020

Rights: Copyright © 2013 INFORMS Applied Probability Society

Vol.3 • No. 2 • 2013
Back to Top