1 / 19

Mathematics vs. “Heuristics”

Heuristics/Experimental Optimisation : When a functional relationship between the variables and the objective function is unknown we experiment on a real object or a scale model. Experiments : Systematic investigation or a strategy ?

sammy
Download Presentation

Mathematics vs. “Heuristics”

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Heuristics/Experimental Optimisation :When a functional relationship between the variables and the objective function is unknown we experiment on a real object or a scale model. Experiments : Systematic investigation or a strategy ? Strategy :Systematic exploitation of available information. Information can be gained during the optimisation process and it should be implemented in the strategy. Mathematics vs. “Heuristics” • Indirect (or analytic) Methods :Attempts to reach the optimum in a single calculation step, without tests or trials. It is based on special mathematical properties of the objective function at the position of an extremum. (A mathematical strategy !) Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

  2. Direct Methods :Solution is approached in a step by step manner (iteratively), at each step (hopefully) improving the value of the objective function. If this cannot be guaranteed proceed by trial and error. Solving Optimisation Problems • Indirect (or analytic) Methods :Attempts to reach the optimum in a single calculation step, without tests or trials. It is based on special mathematical properties of the objective function at the position of an extremum. Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

  3. Analytic Procedures Mac Laurin, 1742 : Necessary and sufficient conditions for a minimum Scheeffer,, 1886 : Proof for multivariable functions. Necessary Condition : Minimize the gradient - system of equations Sufficient conditions : Keep differentiating - in 1D : the lowest order non-vanishing derivative is positive and of even order _-> minimum If the derivative is negative it represents a maximum (if the order is odd we have a saddle point.) In N dimensions : The determinant of the Hessian matrix must be positive as well as the further N-1 subdeterminants of this matrix. Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

  4. When Analytic Procedures go bad • Discontinuity of the objective function and its derivatives • Differentiation may be impossible (e.g. use of a black-box code or experiments) or inaccurate (e.g. noisy data). • Optima can be local or saddle points • Systems of equations (especially non-linear) may be non-soluble or very expensive to solve • ……………… Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

  5. Direct Search Methods (R.Hooke & T.A. Jeeves, 1961) We use the phrase “direct search” to describe sequential examination of trial solutions involving comparison of each trial solution with the “best” obtained up to that time together with a strategy for determining (as a function of earlier results) what the next trial solution will be. The phrase implies our preference, based on experience, for straightforward search strategies, which employ no techniques of classical analysis, except where there is a demonstratable advantage in doing so.” Non-Gradient Based Methods Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

  6. The key feature is that they do not require numerical function values : the relative rank of objective values is sufficient. (Think of skiing - you know the way down but you do not know the exact altitude - and that is all that matters) Direct Search Methods (DSM) Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

  7. Developed by G.E. Box : The key feature is that they do not require numerical function values : the relative rank of objective values is sufficient. (Think of skiing - you know the way down but you do not know the exact altitude - and that is all that matters) EVOP was first used in a process engineering environment, due to shortage of personnel to the dynamic maximization of chemical processes. It was applied in real experiments which sometimes took place over a few years. The EVolutionary OPeration Method Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

  8. The method originally changes two or three parameters at a time (if possible those that have the strongest influence). For two parameters a square is constructed with an initial condition at its center. The corners represent the points in a cycle of trials The corners are tested sequentially, several times if perturbations are acting. The point with the best result becomes the mid-point in the next cycle. (scaling can be changed, as well as choice of points to be taken for the next cycle) The EVolutionary OPeration Method Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

  9. Determine the function value at various points . Declare point with smallest value the minimum. These methods are also called grid methods or tabulation method. Incredibly slow - number of trial inversely proportiona,l to accuracy (but parallel which was not so good when it was invented 1960’s) DSM : Parallel Methods Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

  10. In sequential methods : Trials are made sequentially Intermediate results are used in order to locate the next point They can be classified as : pattern search methods, simplex methods and methods with adaptive sets of search directions. Examples : Boxing the minimum, Interval Division Slow - number of trial proportional to log of accuracy (but sequential which was good when it was invented in 1960’s). DSM : Sequential Methods Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

  11. DSM : Hooke and Jeeves Two types of moves :Exploratory Move - An extrapolation along the line of the first and last move before the variables are varied again individually. Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

  12. As quoted by Davidon (1967) : “ Enrico Fermi and Nicholas Metropolis used one of the first digital computers, the Los Alamos Maniac, to determine which values of certain theoretical parameters best fit experimental data. They varied one theoretical parameter at a time by steps of the same magnitude, and when no such increase or decrease in any one parameter further improved the fit to the experimental data, they halved the step size and repeated the process untill the steps were deemed sufficiently small. This simple procedure was slow, but sure….” Pattern search methods are characterized by a series of exploratory move that consider the behaviour of the objective function at a pattern of points all of which lie on a rationale lattice. The exploratory moves consist of a systematic strategy for visiting the points in the lattice in the immediate vicinity of the current iterate. The key feature is that points remain on a lattice. DSM : Pattern Search Methods Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

  13. Polak (1971) Theorem : If {Xk} is a sequence constructed by the method of pattern search then any accumulation point X* satisfies that Vf(X*) = 0. Key observation : The method can construct only a finite number of intermediate points before reducing the step size by half. Hence the algorithm cannot jam up at a point. Using the lattice property of the method global convergence results can be obtained. Pattern Search Methods Converge Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

  14. Simplex methods were motivated by the fact that earlier DSM’s required from 2n to 2n objective function evaluations to complete the search for an improvement of an iterate. The key observation of Spendley, Hext and Himmworth was that it should not take more than n+1 values for the of the objective to identify a gradient (uphill/downhill) direction. (it makes sense as n+1 points would be needed in order to define a gradient of a function A simplex is a set of n+1 points in Rn. In 2D : a Triangle , In 3D a tetrahedron, etc. DSM : Simplex Search Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

  15. An initial simplex is constructed Identify the vertex with the worst fitness Reflect the worst vertex across the centroid of the opposite The point with the best result becomes the mid-point in the next cycle. ( be careful so as not to keep reflecting back and DSM : Simplex Search Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

  16. With a simplex search : Do we have a new candidate for a minimiser ? Yes, when a reflected vertex produces a strict decrease on the value of the objective of the best vertex. With a simplex search are we near a minimiser ? A circling sequence of simplices could indicate that a neighborhood has been identified. DSM : Simplex Search Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

  17. The simplex can be deformed/adapted so as to take into account the shape of the objective function. DSM : Simplex Search Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

  18. DSM : Rosenbrock’s rotating coords. Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

  19. B contains information for the evolution path - Correlations of successful mutations - PCA of paths I. EVOLUTION STRATEGIES Covariance Matrix Adaptation ES - (N. Hansen) The (1,1) - ES The environment is identified through mutation/success Evolutionary Computation (P. Koumoutsakos)www.icos.ethz.ch

More Related