In some optimization problems the evaluation of the objective function is very expensive. It is therefore desirable to find the global optimum of the function with only comparatively few function evaluations. Because of the expense of evaluations it is justified to put significant effort into finding good sample points and using all the available information about the objective function. One way of achieving this is by assuming that the function can be modelled as a stochastic process and fitting a response surface to it, based on function evaluations at a set of points determined by an initial design. Parameters in the model are estimated when fitting the response surface to the available data. In determining the next point at which to evaluate the objective function, a balance must be struck between local search and global search. Local search in a neighbourhood of the minimum of the approximating function has the aim of finding a point with improved objective value. The aim of global search is to improve the approximation by maximizing an error function which reflects the uncertainty in the approximating function. Such a balance is achieved by using the expected improvement criterion. In this approach the next sample point is chosen where the expected improvement is maximized. The expected improvement at any point in the range reflects the expected amount of improvement of the approximating function beyond a target value (usually the best function value found up to this point) at that point, taking into account the uncertainty in the approximating function. In this thesis, we present and examine the expected improvement approach and the maximization of the expected improvement function.
Identifer | oai:union.ndltd.org:bl.uk/oai:ethos.bl.uk:657484 |
Date | January 2003 |
Creators | Mayer, Theresia |
Publisher | University of Edinburgh |
Source Sets | Ethos UK |
Detected Language | English |
Type | Electronic Thesis or Dissertation |
Source | http://hdl.handle.net/1842/15297 |
Page generated in 0.0013 seconds