The line search approach first finds a descent direction along which the objective function will be reduced and then computes a step size that determines how far should move along that direction. The descent direction can be computed by various methods, such as gradient descent, Newton's method and quasi-Newton method. The step size can be determined either exactly or inexactly.
Here is an example gradient method that uses a line search in step 4.
- Set iteration counter , and make an initial guess for the minimum
- Compute a descent direction
- Choose to 'loosely' minimize over
- Update , and
- Until < tolerance
At the line search step (4) the algorithm might either exactly minimize h, by solving , or loosely, by asking for a sufficient decrease in h. One example of the former is conjugate gradient method. The latter is called inexact line search and may be performed in a number of ways, such as a backtracking line search or using the Wolfe conditions.
Direct search methods
In this method, the minimum must first be bracketed, so the algorithm must identify points x1 and x2 such that the sought minimum lies between them. The interval is then divided by computing at two internal points, x3 and x4, and rejecting whichever of the two outer points is not adjacent to that of x3 and x4 which has the lowest function value. In subsequent steps, only one extra internal point needs to be calculated. Of the various methods of dividing the interval, golden section search is particularly simple and effective, as the interval proportions are preserved regardless of how the search proceeds:
- Backtracking line search
- Secant method
- Newton's method in optimization
- Pattern search (optimization)
- Nelder–Mead method
- Golden section search
- Box, M. J.; Davies, D.; Swann, W. H. (1969). Non-Linear optimisation Techniques. Oliver & Boyd.