Stochastic programming facilitates decision making under uncertainty. It is usually impractical or impossible to find the optimal solution to a stochastic problem, and approximations are required. Sampling-based approximations are simple and attractive, but the standard point estimate of optimization and the Monte Carlo approximation. We provide a method to reduce this bias, and hence provide a better, i.e., tighter, confidence interval on the optimal value and on a candidate solution's optimality gap. Our method requires less restrictive assumptions on the structure of the bias than previously-available estimators. Our estimators adapt to problem-specific properties, and we provide a family of estimators, which allows flexibility in choosing the level of aggressiveness for bias reduction. We establish desirable statistical properties of our estimators and empirically compare them with known techniques on test problems from the literature.
Identifer | oai:union.ndltd.org:UTEXAS/oai:repositories.lib.utexas.edu:2152/3794 |
Date | 29 August 2008 |
Creators | Partani, Amit, 1978- |
Contributors | Morton, David P. |
Source Sets | University of Texas |
Language | English |
Detected Language | English |
Type | Thesis |
Format | electronic |
Rights | Copyright © is held by the author. Presentation of this material on the Libraries' web site by University Libraries, The University of Texas at Austin was made possible under a limited license grant from the author who has retained all copyrights in the works. |
Page generated in 0.0011 seconds