Pattern search (optimization)

From Wikipedia, the free encyclopedia
Jump to navigation Jump to search
Example of convergence of a direct search method on Broyden function

Pattern search (also known as direct search, derivative-free search, or black-box search) is a family of numerical optimization methods that does not require a gradient. As a result, it can be used on functions that are not continuous or differentiable. One such pattern search method is "convergence" (see below), which is based on the theory of positive bases. Optimization attempts to find the best match (the solution that has the lowest error value) in a multidimensional analysis space of possibilities.

History[edit]

The name "pattern search" was coined by Hooke and Jeeves.[1] An early and simple variant is attributed to Fermi and Metropolis when they worked at the Los Alamos National Laboratory. It is described by Davidon,[2] as follows:

They varied one theoretical parameter at a time by steps of the same magnitude, and when no such increase or decrease in any one parameter further improved the fit to the experimental data, they halved the step size and repeated the process until the steps were deemed sufficiently small.

Convergence[edit]

Convergence is a pattern search method proposed by Yu, who proved that it converges using the theory of positive bases.[3] Later, Torczon, Lagarias and co-authors[4][5] used positive-basis techniques to prove the convergence of another pattern-search method on specific classes of functions. Outside of such classes, pattern search is a heuristic that can provide useful approximate solutions for some issues, but can fail on others. Outside of such classes, pattern search is not an iterative method that converges to a solution; indeed, pattern-search methods can converge to non-stationary points on some relatively tame problems.[6][7]

See also[edit]

  • Golden-section search conceptually resembles PS in its narrowing of the search range, only for single-dimensional search spaces.
  • Nelder–Mead method aka. the simplex method conceptually resembles PS in its narrowing of the search range for multi-dimensional search spaces but does so by maintaining n + 1 points for n-dimensional search spaces, whereas PS methods computes 2n + 1 points (the central point and 2 points in each dimension).
  • Luus–Jaakola samples from a uniform distribution surrounding the current position and uses a simple formula for exponentially decreasing the sampling range.
  • Random search is a related family of optimization methods that sample from a hypersphere surrounding the current position.
  • Random optimization is a related family of optimization methods that sample from a normal distribution surrounding the current position.

References[edit]

  1. ^ Hooke, R.; Jeeves, T.A. (1961). ""Direct search" solution of numerical and statistical problems". Journal of the ACM. 8 (2): 212–229. doi:10.1145/321062.321069. 
  2. ^ Davidon, W.C. (1991). "Variable metric method for minimization". SIAM Journal on Optimization. 1 (1): 1–17. doi:10.1137/0801001. 
  3. ^ *Yu, Wen Ci. 1979. “Positive basis and a class of direct search techniques”. Scientia Sinica [Zhongguo Kexue]: 53—68.
    • Yu, Wen Ci. 1979. “The convergent property of the simplex evolutionary technique”. Scientia Sinica [Zhongguo Kexue]: 69–77.
  4. ^ Torczon, V.J. (1997). "On the convergence of pattern search algorithms" (PDF). SIAM Journal on Optimization. 7 (1): 1–25. doi:10.1137/S1052623493250780. 
  5. ^ Dolan, E.D.; Lewis, R.M.; Torczon, V.J. (2003). "On the local convergence of pattern search" (PDF). SIAM Journal on Optimization. 14 (2): 567–583. doi:10.1137/S1052623400374495. 
  6. ^ * Powell, Michael J. D. 1973. ”On Search Directions for Minimization Algorithms.” Mathematical Programming 4: 193—201.
  7. ^ * McKinnon, K. I. M. (1999). "Convergence of the Nelder–Mead simplex method to a non-stationary point". SIAM J. Optimization. 9: 148–158. doi:10.1137/S1052623496303482.  (algorithm summary online).