Incomputational mathematics, aniterative method is amathematical procedure that uses an initial value to generate a sequence of improving approximate solutions for a class of problems, in which thei-th approximation (called an "iterate") is derived from the previous ones.
A specific implementation withtermination criteria for a given iterative method likegradient descent,hill climbing,Newton's method, orquasi-Newton methods likeBFGS, is analgorithm of an iterative method or amethod of successive approximation. An iterative method is calledconvergent if the corresponding sequence converges for given initial approximations. A mathematically rigorous convergence analysis of an iterative method is usually performed; however,heuristic-based iterative methods are also common.
In contrast,direct methods attempt to solve the problem by a finite sequence of operations. In the absence ofrounding errors, direct methods would deliver an exact solution (for example, solving a linear system of equations byGaussian elimination). Iterative methods are often the only choice fornonlinear equations. However, iterative methods are often useful even for linear problems involving many variables (sometimes on the order of millions), where direct methods would be prohibitively expensive (and in some cases impossible) even with the best available computing power.[1]
If an equation can be put into the formf(x) =x, and a solutionx is an attractivefixed point of the functionf, then one may begin with a pointx1 in thebasin of attraction ofx, and letxn+1 =f(xn) forn ≥ 1, and the sequence {xn}n ≥ 1 will converge to the solutionx. Herexn is thenth approximation or iteration ofx andxn+1 is the next orn + 1 iteration ofx. Alternately, superscripts in parentheses are often used in numerical methods, so as not to interfere with subscripts with other meanings. (For example,x(n+1) =f(x(n)).) If the functionf iscontinuously differentiable, a sufficient condition for convergence is that thespectral radius of the derivative is strictly bounded by one in a neighborhood of the fixed point. If this condition holds at the fixed point, then a sufficiently small neighborhood (basin of attraction) must exist.[citation needed]
In the case of asystem of linear equations, the two main classes of iterative methods are thestationary iterative methods, and the more generalKrylov subspace methods.
Stationary iterative methods solve a linear system with anoperator approximating the original one; and based on a measurement of the error in the result (the residual), form a "correction equation" for which this process is repeated. While these methods are simple to derive, implement, and analyze, convergence is only guaranteed for a limited class of matrices.
Aniterative method is defined byand for a given linear system with exact solution theerror byAn iterative method is calledlinear if there exists a matrix such thatand this matrix is called theiteration matrix.An iterative method with a given iteration matrix is calledconvergent if the following holds
An important theorem states that for a given iterative method and its iteration matrix it is convergent if and only if itsspectral radius is smaller than unity, that is,
The basic iterative methods work bysplitting the matrix intoand here the matrix should be easilyinvertible.The iterative methods are now defined asor, equivalently,From this follows that the iteration matrix is given by
Basic examples of stationary iterative methods use a splitting of the matrix such aswhere is only the diagonal part of, and is the strict lowertriangular part of.Respectively, is the strict upper triangular part of.
Linear stationary iterative methods are also calledrelaxation methods.
Krylov subspace methods[2] work by forming abasis of the sequence of successive matrix powers times the initial residual (theKrylov sequence).The approximations to the solution are then formed by minimizing the residual over the subspace formed.The prototypical method in this class is theconjugate gradient method (CG) which assumes that the system matrix issymmetricpositive-definite.For symmetric (and possibly indefinite) one works with theminimal residual method (MINRES).In the case of non-symmetric matrices, methods such as thegeneralized minimal residual method (GMRES) and thebiconjugate gradient method (BiCG) have been derived.
Since these methods form a basis, it is evident that the method converges inN iterations, whereN is the system size. However, in the presence of rounding errors this statement does not hold; moreover, in practiceN can be very large, and the iterative process reaches sufficient accuracy already far earlier. The analysis of these methods is hard, depending on a complicated function of thespectrum of the operator.[citation needed]
The approximating operator that appears in stationary iterative methods can also be incorporated in Krylov subspace methods such asGMRES (alternatively,preconditioned Krylov methods can be considered as accelerations of stationary iterative methods), where they become transformations of the original operator to a presumably better conditioned one. The construction of preconditioners is a large research area.[citation needed]
Mathematical methods relating to successive approximation include:
Jamshīd al-Kāshī used iterative methods to calculate the sine of 1° andπ inThe Treatise of Chord and Sine to high precision.An early iterative method forsolving a linear system appeared in a letter ofGauss to a student of his. He proposed solving a 4-by-4 system of equations by repeatedly solving the component in which the residual was the largest[citation needed].
The theory of stationary iterative methods was solidly established with the work ofD.M. Young starting in the 1950s. The conjugate gradient method was also invented in the 1950s, with independent developments byCornelius Lanczos,Magnus Hestenes andEduard Stiefel, but its nature and applicability were misunderstood at the time. Only in the 1970s was it realized that conjugacy based methods work very well forpartial differential equations, especially the elliptic type.[citation needed]