On the Optimality of (s, S) Policies

 

Eugene A. Feinberg

Department of Applied Mathematics and Statistics

State University of New York at Stony Brook

 

Mark E. Lewis

Cornell University
School of Operations Research and Information Engineering
Ithaca, New York 14853

 

This paper describes results on the existence of optimal policies and convergence properties of optimal actions for discounted and average-cost Markov Decision Processes with weakly continuous transition probabilities. It is possible that cost functions are unbounded and action sets are not compact. The results are applied to stochastic periodic-review inventory control problems, for which they imply the existence of stationary optimal policies and certain optimality properties. The optimality of (s, S) policies is proved by using dynamic programming equations for discounted costs and the vanishing discount factor approach for average costs per unit time.