Jump to content

Successive over-relaxation

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by 130.126.37.192 (talk) at 19:41, 9 September 2009 (External links: link dead, title not found by google). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

In numerical linear algebra, the method of successive over-relaxation (SOR) is a variant of the Gauss–Seidel method for solving a linear system of equations, resulting in faster convergence. A similar method can be used for any slowly converging iterative process. It was devised simultaneously by David M. Young and by H. Frankel in 1950 for the purpose of automatically solving linear systems on digital computers. Over-relaxation methods have been used before the work of Young and Frankel. For instance, the method of Lewis Fry Richardson, and the methods developed by R. V. Southwell. However, these methods were designed for computation by human calculators, and they required some expertise to ensure convergence to the solution which made them inapplicable for programming on digital computers. These aspects are discussed in the thesis of David M. Young.[1]

Formulation

Given a square system of n linear equations with unknown x:

where:

Then A can be decomposed into a diagonal component D, and strictly lower and upper triangular components L and U:

The system of linear equations may be rewritten as:

for a constant ω > 0.

The method of successive over-relaxation is an iterative technique that solves the left hand side of this expression for x, using previous value for x on the right hand side. Analytically, this may be written as:

However, by taking advantage of the triangular form of (D+ωL), the elements of x(k+1) can be computed sequentially using forward substitution:

The choice of relaxation factor is not necessarily easy, and depends upon the properties of the coefficient matrix. For symmetric, positive-definite matrices it can be proven that 0 < ω < 2 will lead to convergence, but we are generally interested in faster convergence rather than just convergence.

Algorithm

Inputs: A , b, ω
Output:

Choose an initial guess to the solution
repeat until convergence

for i from 1 until n do
for j from 1 until i − 1 do
end (j-loop)
for j from i + 1 until n do
end (j-loop)
end (i-loop)
check if convergence is reached

end (repeat)

Note:
can also be written , thus saving one multiplication in each iteration of the outer for-loop.

Other applications of the method

A similar technique can be used for any iterative method. If the original iteration had the form

then the modified version would use

or equivalently

Values of are used to speedup convergence of a slow-converging process, while values of are often used to help establish convergence of a diverging iterative process.

There are various methods that adaptively set the relaxation parameter based on the observed behavior of the converging process. Usually they help to reach a super-linear convergence for some problems but fail for the others

See also

Notes

  1. ^ Young, David M. (1950), Iterative methods for solving partial difference equations of elliptical type (PDF), PhD thesis, Harvard University, retrieved 2009-06-15 {{citation}}: Unknown parameter |month= ignored (help)

References