Jump to content

Successive over-relaxation: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
seperated a paragraph about the convergence into its own section
Convergence: added the convergence theorem
Line 44: Line 44:
In 1947, Ostrowski proved that if <math>A</math> is [[Symmetric matrix|symmetric]] and [[Positive-definite matrix|positive-definite]] then <math>\rho(L_\omega)<1</math> for <math>0<\omega<2 </math>.
In 1947, Ostrowski proved that if <math>A</math> is [[Symmetric matrix|symmetric]] and [[Positive-definite matrix|positive-definite]] then <math>\rho(L_\omega)<1</math> for <math>0<\omega<2 </math>.
Thus, convergence of the iteration process follows, but we are generally interested in faster convergence rather than just convergence.
Thus, convergence of the iteration process follows, but we are generally interested in faster convergence rather than just convergence.

=== Convergence Rate ===
The convergence rate for the SOR method can be analytically derived.
One needs to assume the following
* the relaxation parameter is appropriate: <math> \omega \in (0,2) </math>
* [[Jacobi_method|Jacobi's]] iteration matrix <math> C_\text{Jac}:= I-D^{-1}A </math> has only real eigenvalues
* [[Jacobi_method|Jacobi's method]] is convergent: <math> \mu := \rho(C_\text{Jac}) < 1 </math>
* a unique solution exists: <math> \det A \neq 0 </math>.
Then the convergence rate can be expressed as<ref>{{Cite book|url=https://doi.org/10.1007/978-3-319-28483-5|title=Iterative Solution of Large Sparse Systems of Equations {{!}} SpringerLink|last=Hackbusch|first=Wolfgang|publisher=|year=|isbn=|location=|pages=|language=en-gb|chapter=4.6.2|doi=10.1007/978-3-319-28483-5}}</ref>
:<math>
\rho(C_\omega) =
\begin{cases}
\frac{1}{4} \left( \omega \mu + \sqrt{\omega^2 \mu^2-4(\omega-1)} \right)^2\,,
& 0 < \omega \leq \omega_\text{opt}
\\
\omega -1\,,
& \omega_\text{opt} < \omega < 2
\end{cases}
</math>
where the optimal relaxation parameter is given by
:<math>
\omega_\text{opt} := 1+ \left( \frac{\mu}{1+\sqrt{1-\mu^2}} \right)^2\,.
</math>
[[File:Spectral Radius.svg|frame|Spectral radius <math> \rho(C_\omega) </math> of the iteration matrix for the SOR method <math> C_\omega </math>.
The plot shows the dependence on the spectral radius of the Jacobi iteration matrix <math> \mu := \rho(C_\text{Jac}) </math>.]]


== Algorithm ==
== Algorithm ==

Revision as of 14:04, 26 March 2018

In numerical linear algebra, the method of successive over-relaxation (SOR) is a variant of the Gauss–Seidel method for solving a linear system of equations, resulting in faster convergence. A similar method can be used for any slowly converging iterative process.

It was devised simultaneously by David M. Young, Jr. and by Stanley P. Frankel in 1950 for the purpose of automatically solving linear systems on digital computers. Over-relaxation methods had been used before the work of Young and Frankel. An example is the method of Lewis Fry Richardson, and the methods developed by R. V. Southwell. However, these methods were designed for computation by human calculators, and they required some expertise to ensure convergence to the solution which made them inapplicable for programming on digital computers. These aspects are discussed in the thesis of David M. Young, Jr.[1]

Formulation

Given a square system of n linear equations with unknown x:

where:

Then A can be decomposed into a diagonal component D, and strictly lower and upper triangular components L and U:

where

The system of linear equations may be rewritten as:

for a constant ω > 1, called the relaxation factor.

The method of successive over-relaxation is an iterative technique that solves the left hand side of this expression for x, using previous value for x on the right hand side. Analytically, this may be written as:

where is the kth approximation or iteration of and is the next or k + 1 iteration of . However, by taking advantage of the triangular form of (D+ωL), the elements of x(k+1) can be computed sequentially using forward substitution:

Convergence

The choice of relaxation factor ω is not necessarily easy, and depends upon the properties of the coefficient matrix. In 1947, Ostrowski proved that if is symmetric and positive-definite then for . Thus, convergence of the iteration process follows, but we are generally interested in faster convergence rather than just convergence.

Convergence Rate

The convergence rate for the SOR method can be analytically derived. One needs to assume the following

  • the relaxation parameter is appropriate:
  • Jacobi's iteration matrix has only real eigenvalues
  • Jacobi's method is convergent:
  • a unique solution exists: .

Then the convergence rate can be expressed as[2]

where the optimal relaxation parameter is given by

Spectral radius of the iteration matrix for the SOR method . The plot shows the dependence on the spectral radius of the Jacobi iteration matrix .

Algorithm

Since elements can be overwritten as they are computed in this algorithm, only one storage vector is needed, and vector indexing is omitted. The algorithm goes as follows:

Inputs: A, b, ω
Output: 
Choose an initial guess  to the solution
repeat until convergence
  for i from 1 until n do
    
    for j from 1 until n do
      if ji then
        
      end if
    end (j-loop)
    
  end (i-loop)
  check if convergence is reached
end (repeat)
Note
can also be written , thus saving one multiplication in each iteration of the outer for-loop.

Symmetric successive over-relaxation

The version for symmetric matrices A, in which

is referred to as Symmetric Successive Over-Relaxation, or (SSOR), in which

and the iterative method is

The SOR and SSOR methods are credited to David M. Young, Jr..

Other applications of the method

A similar technique can be used for any iterative method. If the original iteration had the form

then the modified version would use

Note however that the formulation presented above, used for solving systems of linear equations, is not a special case of this formulation if x is considered to be the complete vector. If this formulation is used instead, the equation for calculating the next vector will look like

where . Values of are used to speed up convergence of a slow-converging process, while values of are often used to help establish convergence of a diverging iterative process or speed up the convergence of an overshooting process.

There are various methods that adaptively set the relaxation parameter based on the observed behavior of the converging process. Usually they help to reach a super-linear convergence for some problems but fail for the others.

See also

Notes

  1. ^ Young, David M. (May 1, 1950), Iterative methods for solving partial difference equations of elliptical type (PDF), PhD thesis, Harvard University, retrieved 2009-06-15
  2. ^ Hackbusch, Wolfgang. "4.6.2". Iterative Solution of Large Sparse Systems of Equations | SpringerLink. doi:10.1007/978-3-319-28483-5.

References