Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Eigenvalue algorithm

From Wikipedia, the free encyclopedia
Numerical methods for matrix eigenvalue calculation

Innumerical analysis, one of the most important problems is designing efficient andstablealgorithms for finding theeigenvalues of amatrix. Theseeigenvalue algorithms may also find eigenvectors.

Eigenvalues and eigenvectors

[edit]
Main articles:Eigenvalues and eigenvectors andGeneralized eigenvector

Given ann ×nsquare matrixA ofreal orcomplex numbers, aneigenvalueλ and its associatedgeneralized eigenvectorv are a pair obeying the relation[1]

(AλI)kv=0,{\displaystyle \left(A-\lambda I\right)^{k}{\mathbf {v} }=0,}

wherev is a nonzeron × 1 column vector,I is then ×nidentity matrix,k is a positive integer, and bothλ andv are allowed to be complex even whenA is real. Whenk = 1, the vector is called simply aneigenvector, and the pair is called aneigenpair. In this case,Av =λv. Any eigenvalueλ ofA has ordinary[note 1] eigenvectors associated to it, for ifk is the smallest integer such that(AλI)kv = 0 for a generalized eigenvectorv, then(AλI)k−1v is an ordinary eigenvector. The valuek can always be taken as less than or equal ton. In particular,(AλI)nv = 0 for all generalized eigenvectorsv associated withλ.

For each eigenvalueλ ofA, thekernelker(AλI) consists of all eigenvectors associated withλ (along with 0), called theeigenspace ofλ, while the vector spaceker((AλI)n) consists of all generalized eigenvectors, and is called thegeneralized eigenspace. Thegeometric multiplicity ofλ is the dimension of its eigenspace. Thealgebraic multiplicity ofλ is the dimension of its generalized eigenspace. The latter terminology is justified by the equation

pA(z)=det(zIA)=i=1k(zλi)αi,{\displaystyle p_{A}\left(z\right)=\det \left(zI-A\right)=\prod _{i=1}^{k}(z-\lambda _{i})^{\alpha _{i}},}

wheredet is thedeterminant function, theλi are all the distinct eigenvalues ofA and theαi are the corresponding algebraic multiplicities. The functionpA(z) is thecharacteristic polynomial ofA. So the algebraic multiplicity is the multiplicity of the eigenvalue as azero of the characteristic polynomial. Since any eigenvector is also a generalized eigenvector, the geometric multiplicity is less than or equal to the algebraic multiplicity. The algebraic multiplicities sum up ton, the degree of the characteristic polynomial. The equationpA(z) = 0 is called thecharacteristic equation, as its roots are exactly the eigenvalues ofA. By theCayley–Hamilton theorem,A itself obeys the same equation:pA(A) = 0.[note 2] As a consequence, the columns of the matrixij(AλiI)αi{\textstyle \prod _{i\neq j}(A-\lambda _{i}I)^{\alpha _{i}}} must be either 0 or generalized eigenvectors of the eigenvalueλj, since they are annihilated by(AλjI)αj{\displaystyle (A-\lambda _{j}I)^{\alpha _{j}}}. In fact, thecolumn space is the generalized eigenspace ofλj.

Any collection of generalized eigenvectors of distinct eigenvalues is linearly independent, so a basis for all ofCn can be chosen consisting of generalized eigenvectors. More particularly, this basis{vi}n
i=1
can be chosen and organized so that

  • ifvi andvj have the same eigenvalue, then so doesvk for eachk betweeni andj, and
  • ifvi is not an ordinary eigenvector, and ifλi is its eigenvalue, then(AλiI)vi =vi−1 (in particular,v1 must be an ordinary eigenvector).

If these basis vectors are placed as the column vectors of a matrixV = [v1v2vn], thenV can be used to convertA to itsJordan normal form:

V1AV=[λ1β1000λ2β2000λ30000λn],{\displaystyle V^{-1}AV={\begin{bmatrix}\lambda _{1}&\beta _{1}&0&\ldots &0\\0&\lambda _{2}&\beta _{2}&\ldots &0\\0&0&\lambda _{3}&\ldots &0\\\vdots &\vdots &\vdots &\ddots &\vdots \\0&0&0&\ldots &\lambda _{n}\end{bmatrix}},}

where theλi are the eigenvalues,βi = 1 if(Aλi+1)vi+1 =vi andβi = 0 otherwise.

More generally, ifW is any invertible matrix, andλ is an eigenvalue ofA with generalized eigenvectorv, then(W−1AWλI)kWkv = 0. Thusλ is an eigenvalue ofW−1AW with generalized eigenvectorWkv. That is,similar matrices have the same eigenvalues.

Normal, Hermitian, and real-symmetric matrices

[edit]
Main articles:Adjoint matrix,Normal matrix, andHermitian matrix

TheadjointM* of a complex matrixM is the transpose of the conjugate ofM:M* =MT. A square matrixA is callednormal if it commutes with its adjoint:A*A =AA*. It is calledHermitian if it is equal to its adjoint:A* =A. All Hermitian matrices are normal. IfA has only real elements, then the adjoint is just the transpose, andA is Hermitian if and only if it issymmetric. When applied to column vectors, the adjoint can be used to define the canonical inner product onCn:wv =w*v.[note 3] Normal, Hermitian, and real-symmetric matrices have several useful properties:

  • Every generalized eigenvector of a normal matrix is an ordinary eigenvector.
  • Any normal matrix is similar to a diagonal matrix, since its Jordan normal form is diagonal.
  • Eigenvectors of distinct eigenvalues of a normal matrix are orthogonal.
  • The null space and the image (or column space) of a normal matrix are orthogonal to each other.
  • For any normal matrixA,Cn has an orthonormal basis consisting of eigenvectors ofA. The corresponding matrix of eigenvectors isunitary.
  • The eigenvalues of a Hermitian matrix are real, since(λλ)v = (A*A)v = (AA)v = 0 for a non-zero eigenvectorv.
  • IfA is real, there is an orthonormal basis forRn consisting of eigenvectors ofA if and only ifA is symmetric.

It is possible for a real or complex matrix to have all real eigenvalues without being Hermitian. For example, a realtriangular matrix has its eigenvalues along its diagonal, but in general is not symmetric.

Condition number

[edit]

Any problem of numeric calculation can be viewed as the evaluation of some functionf for some inputx. Thecondition numberκ(f,x) of the problem is the ratio of the relative error in the function's output to the relative error in the input, and varies with both the function and the input. The condition number describes how error grows during the calculation. Its base-10 logarithm tells how many fewer digits of accuracy exist in the result than existed in the input. The condition number is a best-case scenario. It reflects the instability built into the problem, regardless of how it is solved. No algorithm can ever produce more accurate results than indicated by the condition number, except by chance. However, a poorly designed algorithm may produce significantly worse results. For example, as mentioned below, the problem of finding eigenvalues for normal matrices is always well-conditioned. However, the problem of finding the roots of a polynomial can bevery ill-conditioned. Thus eigenvalue algorithms that work by finding the roots of the characteristic polynomial can be ill-conditioned even when the problem is not.

For the problem of solving the linear equationAv =b whereA is invertible, thematrix condition numberκ(A−1,b) is given by||A||op||A−1||op, where|| ||op is theoperator norm subordinate to the normalEuclidean norm onCn. Since this number is independent ofb and is the same forA andA−1, it is usually just called the condition numberκ(A) of the matrixA. This valueκ(A) is also the absolute value of the ratio of the largestsingular value ofA to its smallest. IfA isunitary, then||A||op = ||A−1||op = 1, soκ(A) = 1. For general matrices, the operator norm is often difficult to calculate. For this reason, othermatrix norms are commonly used to estimate the condition number.

For the eigenvalue problem,Bauer and Fike proved that ifλ is an eigenvalue for adiagonalizablen ×n matrixA witheigenvector matrixV, then the absolute error in calculatingλ is bounded by the product ofκ(V) and the absolute error inA.[2]As a result, the condition number for findingλ isκ(λ,A) =κ(V) = ||V ||op ||V−1||op. IfA is normal, thenV is unitary, andκ(λ,A) = 1. Thus the eigenvalue problem for all normal matrices is well-conditioned.

The condition number for the problem of finding the eigenspace of a normal matrixA corresponding to an eigenvalueλ has been shown to be inversely proportional to the minimum distance betweenλ and the other distinct eigenvalues ofA.[3] In particular, the eigenspace problem for normal matrices is well-conditioned for isolated eigenvalues. When eigenvalues are not isolated, the best that can be hoped for is to identify the span of all eigenvectors of nearby eigenvalues.

Algorithms

[edit]

The most reliable and most widely used algorithm for computing eigenvalues isJohn G. F. Francis' andVera N. Kublanovskaya'sQR algorithm, considered one of the top ten algorithms of 20th century.[4]

Any monic polynomial is the characteristic polynomial of itscompanion matrix. Therefore, a general algorithm for finding eigenvalues could also be used to find the roots of polynomials. TheAbel–Ruffini theorem shows that any such algorithm for dimensions greater than 4 must either be infinite, or involve functions of greater complexity than elementary arithmetic operations and fractional powers. For this reason algorithms that exactly calculate eigenvalues in a finite number of steps only exist for a few special classes of matrices. For general matrices, algorithms areiterative, producing better approximate solutions with each iteration.

Some algorithms produce every eigenvalue, others will produce a few, or only one. However, even the latter algorithms can be used to find all eigenvalues. Once an eigenvalueλ of a matrixA has been identified, it can be used to either direct the algorithm towards a different solution next time, or to reduce the problem to one that no longer hasλ as a solution.

Redirection is usually accomplished by shifting: replacingA withAμI for some constantμ. The eigenvalue found forAμI must haveμ added back in to get an eigenvalue forA. For example, forpower iteration,μ =λ. Power iteration finds the largest eigenvalue in absolute value, so even whenλ is only an approximate eigenvalue, power iteration is unlikely to find it a second time. Conversely,inverse iteration based methods find the lowest eigenvalue, soμ is chosen well away fromλ and hopefully closer to some other eigenvalue.

Reduction can be accomplished by restrictingA to the column space of the matrixAλI, whichA carries to itself. SinceA -λI is singular, the column space is of lesser dimension. The eigenvalue algorithm can then be applied to the restricted matrix. This process can be repeated until all eigenvalues are found.

If an eigenvalue algorithm does not produce eigenvectors, a common practice is to use an inverse iteration based algorithm withμ set to a close approximation to the eigenvalue. This will quickly converge to the eigenvector of the closest eigenvalue toμ. For small matrices, an alternative is to look at the column space of the product ofAλ'I for each of the other eigenvaluesλ'.

A formula for the norm of unit eigenvector components of normal matrices was discovered by Robert Thompson in 1966 and rediscovered independently by several others.[5][6][7][8][9]IfA is ann×n{\textstyle n\times n} normal matrix with eigenvaluesλi(A) and corresponding unit eigenvectorsvi whose component entries arevi,j, letAj be then1×n1{\textstyle n-1\times n-1} matrix obtained by removing thei-th row and column fromA, and letλk(Aj) be itsk-th eigenvalue. Then|vi,j|2k=1,kin(λi(A)λk(A))=k=1n1(λi(A)λk(Aj)){\displaystyle |v_{i,j}|^{2}\prod _{k=1,k\neq i}^{n}(\lambda _{i}(A)-\lambda _{k}(A))=\prod _{k=1}^{n-1}(\lambda _{i}(A)-\lambda _{k}(A_{j}))}

Ifp,pj{\displaystyle p,p_{j}} are the characteristic polynomials ofA{\displaystyle A} andAj{\displaystyle A_{j}}, the formula can be re-written as|vi,j|2=pj(λi(A))p(λi(A)){\displaystyle |v_{i,j}|^{2}={\frac {p_{j}(\lambda _{i}(A))}{p'(\lambda _{i}(A))}}}assuming the derivativep{\displaystyle p'} is not zero atλi(A){\displaystyle \lambda _{i}(A)}.

Hessenberg and tridiagonal matrices

[edit]
Main article:Hessenberg matrix

Because the eigenvalues of a triangular matrix are its diagonal elements, for general matrices there is no finite method likegaussian elimination to convert a matrix to triangular form while preserving eigenvalues. But it is possible to reach something close to triangular. Anupper Hessenberg matrix is a square matrix for which all entries below thesubdiagonal are zero. A lower Hessenberg matrix is one for which all entries above thesuperdiagonal are zero. Matrices that are both upper and lower Hessenberg aretridiagonal. Hessenberg and tridiagonal matrices are the starting points for many eigenvalue algorithms because the zero entries reduce the complexity of the problem. Several methods are commonly used to convert a general matrix into a Hessenberg matrix with the same eigenvalues. If the original matrix was symmetric or Hermitian, then the resulting matrix will be tridiagonal.

When only eigenvalues are needed, there is no need to calculate the similarity matrix, as the transformed matrix has the same eigenvalues. If eigenvectors are needed as well, the similarity matrix may be needed to transform the eigenvectors of the Hessenberg matrix back into eigenvectors of the original matrix.

MethodApplies toProducesCost without similarity matrixCost with similarity matrixDescription
Householder transformationsGeneralHessenberg2n33 +O(n2)[10]: 474 4n33 +O(n2)[10]: 474 Reflect each column through a subspace to zero out its lower entries.
Givens rotationsGeneralHessenberg4n33 +O(n2)[10]: 470 Apply planar rotations to zero out individual entries. Rotations are ordered so that later ones do not cause zero entries to become non-zero again.
Arnoldi iterationGeneralHessenbergPerform Gram–Schmidt orthogonalization on Krylov subspaces.
Lanczos algorithmHermitianTridiagonalArnoldi iteration for Hermitian matrices, with shortcuts.

For symmetric tridiagonal eigenvalue problems all eigenvalues (without eigenvectors) can be computed numerically in time O(n log(n)), using bisection on the characteristic polynomial.[11]

Iterative algorithms

[edit]

Iterative algorithms solve the eigenvalue problem by producing sequences that converge to the eigenvalues. Some algorithms also produce sequences of vectors that converge to the eigenvectors. Most commonly, the eigenvalue sequences are expressed as sequences of similar matrices which converge to a triangular or diagonal form, allowing the eigenvalues to be read easily. The eigenvector sequences are expressed as the corresponding similarity matrices.

MethodApplies toProducesCost per stepConvergenceDescription
Lanczos algorithmHermitianm largest/smallest eigenpairs
Power iterationgeneraleigenpair with largest valueO(n2)linearRepeatedly applies the matrix to an arbitrary starting vector and renormalizes.
Inverse iterationgeneraleigenpair with value closest toμlinearPower iteration for(AμI)−1
Rayleigh quotient iterationHermitianany eigenpaircubicPower iteration for(AμiI)−1, whereμi for each iteration is the Rayleigh quotient of the previous iteration.
Preconditioned inverse iteration[12] orLOBPCG algorithmpositive-definite real symmetriceigenpair with value closest toμInverse iteration using apreconditioner (an approximate inverse toA).
Bisection methodreal symmetric tridiagonalany eigenvaluelinearUses thebisection method to find roots of the characteristic polynomial, supported by the Sturm sequence.
Laguerre iterationreal symmetric tridiagonalany eigenvaluecubic[13]UsesLaguerre's method to find roots of the characteristic polynomial, supported by the Sturm sequence.
QR algorithmHessenbergall eigenvaluesO(n2)cubicFactorsA =QR, whereQ is orthogonal andR is triangular, then applies the next iteration toRQ.
all eigenpairs6n3 +O(n2)
Jacobi eigenvalue algorithmreal symmetricall eigenvaluesO(n3)quadraticUses Givens rotations to attempt clearing all off-diagonal entries. This fails, but strengthens the diagonal.
Divide-and-conquerHermitian tridiagonalall eigenvaluesO(n2)Divides the matrix into submatrices that are diagonalized then recombined.
all eigenpairs(43)n3 +O(n2)
Homotopy methodreal symmetric tridiagonalall eigenpairsO(n2)[14]Constructs a computable homotopy path from a diagonal eigenvalue problem.
Folded spectrum methodreal symmetriceigenpair with value closest toμPreconditioned inverse iteration applied to(AμI)2
MRRR algorithm[15]real symmetric tridiagonalsome or all eigenpairsO(n2)"Multiple relatively robust representations" – performs inverse iteration on aLDLT decomposition of the shifted matrix.
Gram iteration[16]generalEigenpair with largest eigenvaluesuper-linearRepeatedly computes the Gram product and rescales, deterministically.

Direct calculation

[edit]

While there is no simple algorithm to directly calculate eigenvalues for general matrices, there are numerous special classes of matrices where eigenvalues can be directly calculated. These include:

Triangular matrices

[edit]

Since the determinant of atriangular matrix is the product of its diagonal entries, ifT is triangular, thendet(λIT)=i(λTii){\textstyle \det(\lambda I-T)=\prod _{i}(\lambda -T_{ii})}. Thus the eigenvalues ofT are its diagonal entries.

Factorable polynomial equations

[edit]

Ifp is any polynomial andp(A) = 0, then the eigenvalues ofA also satisfy the same equation. Ifp happens to have a known factorization, then the eigenvalues ofA lie among its roots.

For example, aprojection is a square matrixP satisfyingP2 =P. The roots of the corresponding scalar polynomial equation,λ2 =λ, are 0 and 1. Thus any projection has 0 and 1 for its eigenvalues. The multiplicity of 0 as an eigenvalue is thenullity ofP, while the multiplicity of 1 is the rank ofP.

Another example is a matrixA that satisfiesA2 =α2I for some scalarα. The eigenvalues must be±α. The projection operators

P+=12(I+Aα){\displaystyle P_{+}={\frac {1}{2}}\left(I+{\frac {A}{\alpha }}\right)}
P=12(IAα){\displaystyle P_{-}={\frac {1}{2}}\left(I-{\frac {A}{\alpha }}\right)}

satisfy

AP+=αP+AP=αP{\displaystyle AP_{+}=\alpha P_{+}\quad AP_{-}=-\alpha P_{-}}

and

P+P+=P+PP=PP+P=PP+=0.{\displaystyle P_{+}P_{+}=P_{+}\quad P_{-}P_{-}=P_{-}\quad P_{+}P_{-}=P_{-}P_{+}=0.}

Thecolumn spaces ofP+ andP are the eigenspaces ofA corresponding to+α andα, respectively.

2×2 matrices

[edit]

For dimensions 2 through 4, formulas involving radicals exist that can be used to find the eigenvalues. While a common practice for 2×2 and 3×3 matrices, for 4×4 matrices the increasing complexity of theroot formulas makes this approach less attractive.

For the 2×2 matrix

A=[abcd],{\displaystyle A={\begin{bmatrix}a&b\\c&d\end{bmatrix}},}

the characteristic polynomial is

det[λabcλd]=λ2(a+d)λ+(adbc)=λ2λtr(A)+det(A).{\displaystyle \det {\begin{bmatrix}\lambda -a&-b\\-c&\lambda -d\end{bmatrix}}=\lambda ^{2}\,-\,\left(a+d\right)\lambda \,+\,\left(ad-bc\right)=\lambda ^{2}\,-\,\lambda \,{\rm {tr}}(A)\,+\,\det(A).}

Thus the eigenvalues can be found by using thequadratic formula:

λ=tr(A)±tr2(A)4det(A)2.{\displaystyle \lambda ={\frac {{\rm {tr}}(A)\pm {\sqrt {{\rm {tr}}^{2}(A)-4\det(A)}}}{2}}.}

Defininggap(A)=tr2(A)4det(A){\textstyle {\rm {gap}}\left(A\right)={\sqrt {{\rm {tr}}^{2}(A)-4\det(A)}}} to be the distance between the two eigenvalues, it is straightforward to calculate

λa=12(1±adgap(A)),λb=±cgap(A){\displaystyle {\frac {\partial \lambda }{\partial a}}={\frac {1}{2}}\left(1\pm {\frac {a-d}{{\rm {gap}}(A)}}\right),\qquad {\frac {\partial \lambda }{\partial b}}={\frac {\pm c}{{\rm {gap}}(A)}}}

with similar formulas forc andd. From this it follows that the calculation is well-conditioned if the eigenvalues are isolated.

Eigenvectors can be found by exploiting theCayley–Hamilton theorem. Ifλ1,λ2 are the eigenvalues, then(Aλ1I)(Aλ2I) = (Aλ2I)(Aλ1I) = 0, so the columns of(Aλ2I) are annihilated by(Aλ1I) and vice versa. Assuming neither matrix is zero, the columns of each must include eigenvectors for the other eigenvalue. (If either matrix is zero, thenA is a multiple of the identity and any non-zero vector is an eigenvector.)

For example, suppose

A=[4323],{\displaystyle A={\begin{bmatrix}4&3\\-2&-3\end{bmatrix}},}

thentr(A) = 4 − 3 = 1 anddet(A) = 4(−3) − 3(−2) = −6, so the characteristic equation is

0=λ2λ6=(λ3)(λ+2),{\displaystyle 0=\lambda ^{2}-\lambda -6=(\lambda -3)(\lambda +2),}

and the eigenvalues are 3 and -2. Now,

A3I=[1326],A+2I=[6321].{\displaystyle A-3I={\begin{bmatrix}1&3\\-2&-6\end{bmatrix}},\qquad A+2I={\begin{bmatrix}6&3\\-2&-1\end{bmatrix}}.}

In both matrices, the columns are multiples of each other, so either column can be used. Thus,(1, −2) can be taken as an eigenvector associated with the eigenvalue -2, and(3, −1) as an eigenvector associated with the eigenvalue 3, as can be verified by multiplying them byA.

Symmetric 3×3 matrices

[edit]

The characteristic equation of a symmetric 3×3 matrixA is:

det(αIA)=α3α2tr(A)α12(tr(A2)tr2(A))det(A)=0.{\displaystyle \det \left(\alpha I-A\right)=\alpha ^{3}-\alpha ^{2}{\rm {tr}}(A)-\alpha {\frac {1}{2}}\left({\rm {tr}}(A^{2})-{\rm {tr}}^{2}(A)\right)-\det(A)=0.}

This equation may be solved using the methods ofCardano orLagrange, but an affine change toA will simplify the expression considerably, and lead directly to atrigonometric solution. IfA =pB +qI, thenA andB have the same eigenvectors, andβ is an eigenvalue ofB if and only ifα = +q is an eigenvalue ofA. Lettingq=tr(A)/3{\textstyle q={\rm {tr}}(A)/3} andp=(tr((AqI)2)/6)1/2{\textstyle p=\left({\rm {tr}}\left((A-qI)^{2}\right)/6\right)^{1/2}}, gives

det(βIB)=β33βdet(B)=0.{\displaystyle \det \left(\beta I-B\right)=\beta ^{3}-3\beta -\det(B)=0.}

The substitutionβ = 2cosθ and some simplification using the identitycos 3θ = 4cos3θ − 3cosθ reduces the equation tocos 3θ = det(B) / 2. Thus

β=2cos(13arccos(det(B)/2)+2kπ3),k=0,1,2.{\displaystyle \beta =2{\cos }\left({\frac {1}{3}}{\arccos }\left(\det(B)/2\right)+{\frac {2k\pi }{3}}\right),\quad k=0,1,2.}

Ifdet(B) is complex or is greater than 2 in absolute value, the arccosine should be taken along the same branch for all three values ofk. This issue doesn't arise whenA is real and symmetric, resulting in a simple algorithm:[17]

% Given a real symmetric 3x3 matrix A, compute the eigenvalues% Note that acos and cos operate on angles in radiansp1=A(1,2)^2+A(1,3)^2+A(2,3)^2if(p1==0)% A is diagonal.eig1=A(1,1)eig2=A(2,2)eig3=A(3,3)elseq=trace(A)/3% trace(A) is the sum of all diagonal valuesp2=(A(1,1)-q)^2+(A(2,2)-q)^2+(A(3,3)-q)^2+2*p1p=sqrt(p2/6)B=(1/p)*(A-q*I)% I is the identity matrixr=det(B)/2% In exact arithmetic for a symmetric matrix  -1 <= r <= 1% but computation error can leave it slightly outside this range.if(r<=-1)phi=pi/3elseif(r>=1)phi=0elsephi=acos(r)/3end% the eigenvalues satisfy eig3 <= eig2 <= eig1eig1=q+2*p*cos(phi)eig3=q+2*p*cos(phi+(2*pi/3))eig2=3*q-eig1-eig3% since trace(A) = eig1 + eig2 + eig3end

Once again, the eigenvectors ofA can be obtained by recourse to theCayley–Hamilton theorem. Ifα1,α2,α3 are distinct eigenvalues ofA, then(Aα1I)(Aα2I)(Aα3I) = 0. Thus the columns of the product of any two of these matrices will contain an eigenvector for the third eigenvalue. However, ifα3 =α1, then(Aα1I)2(Aα2I) = 0 and(Aα2I)(Aα1I)2 = 0. Thus thegeneralized eigenspace ofα1 is spanned by the columns ofAα2I while the ordinary eigenspace is spanned by the columns of(Aα1I)(Aα2I). The ordinary eigenspace ofα2 is spanned by the columns of(Aα1I)2.

For example, let

A=[326225214].{\displaystyle A={\begin{bmatrix}3&2&6\\2&2&5\\-2&-1&-4\end{bmatrix}}.}

The characteristic equation is

0=λ3λ2λ+1=(λ1)2(λ+1),{\displaystyle 0=\lambda ^{3}-\lambda ^{2}-\lambda +1=(\lambda -1)^{2}(\lambda +1),}

with eigenvalues 1 (of multiplicity 2) and -1. Calculating,

AI=[226215215],A+I=[426235213]{\displaystyle A-I={\begin{bmatrix}2&2&6\\2&1&5\\-2&-1&-5\end{bmatrix}},\qquad A+I={\begin{bmatrix}4&2&6\\2&3&5\\-2&-1&-3\end{bmatrix}}}

and

(AI)2=[408408408],(AI)(A+I)=[044022022]{\displaystyle (A-I)^{2}={\begin{bmatrix}-4&0&-8\\-4&0&-8\\4&0&8\end{bmatrix}},\qquad (A-I)(A+I)={\begin{bmatrix}0&4&4\\0&2&2\\0&-2&-2\end{bmatrix}}}

Thus(−4, −4, 4) is an eigenvector for −1, and(4, 2, −2) is an eigenvector for 1.(2, 3, −1) and(6, 5, −3) are both generalized eigenvectors associated with 1, either one of which could be combined with(−4, −4, 4) and(4, 2, −2) to form a basis of generalized eigenvectors ofA. Once found, the eigenvectors can be normalized if needed.

Eigenvectors of normal 3×3 matrices

[edit]

If a 3×3 matrixA{\displaystyle A} is normal, then the cross-product can be used to find eigenvectors. Ifλ{\displaystyle \lambda } is an eigenvalue ofA{\displaystyle A}, then the null space ofAλI{\displaystyle A-\lambda I} is perpendicular to its column space. Thecross product of two independent columns ofAλI{\displaystyle A-\lambda I} will be in the null space. That is, it will be an eigenvector associated withλ{\displaystyle \lambda }. Since the column space is two dimensional in this case, the eigenspace must be one dimensional, so any other eigenvector will be parallel to it.

IfAλI{\displaystyle A-\lambda I} does not contain two independent columns but is not0, the cross-product can still be used. In this caseλ{\displaystyle \lambda } is an eigenvalue of multiplicity 2, so any vector perpendicular to the column space will be an eigenvector. Supposev{\displaystyle \mathbf {v} } is a non-zero column ofAλI{\displaystyle A-\lambda I}. Choose an arbitrary vectoru{\displaystyle \mathbf {u} } not parallel tov{\displaystyle \mathbf {v} }. Thenv×u{\displaystyle \mathbf {v} \times \mathbf {u} } and(v×u)×v{\displaystyle (\mathbf {v} \times \mathbf {u} )\times \mathbf {v} } will be perpendicular tov{\displaystyle \mathbf {v} } and thus will be eigenvectors ofλ{\displaystyle \lambda }.

This does not work whenA{\displaystyle A} is not normal, as the null space and column space do not need to be perpendicular for such matrices.

See also

[edit]

Notes

[edit]
  1. ^The term "ordinary" is used here only to emphasize the distinction between "eigenvector" and "generalized eigenvector".
  2. ^where the constant term is multiplied by the identity matrixI.
  3. ^This ordering of the inner product (with the conjugate-linear position on the left), is preferred by physicists. Algebraists often place the conjugate-linear position on the right:wv =v*w.

References

[edit]
  1. ^Axler, Sheldon (1995),"Down with Determinants!"(PDF),American Mathematical Monthly,102 (2):139–154,doi:10.2307/2975348,JSTOR 2975348, archived fromthe original(PDF) on 2012-09-13, retrieved2012-07-31
  2. ^F. L. Bauer; C. T. Fike (1960), "Norms and exclusion theorems",Numer. Math.,2:137–141,doi:10.1007/bf01386217,S2CID 121278235
  3. ^S.C. Eisenstat; I.C.F. Ipsen (1998),"Relative Perturbation Results for Eigenvalues and Eigenvectors of Diagonalisable Matrices",BIT,38 (3):502–9,doi:10.1007/bf02510256,S2CID 119886389
  4. ^J. Dongarra and F. Sullivan (2000). "Top ten algorithms of the century".Computing in Science and Engineering.2: 22-23.doi:10.1109/MCISE.2000.814652.
  5. ^Thompson, R. C. (June 1966)."Principal submatrices of normal and Hermitian matrices".Illinois Journal of Mathematics.10 (2):296–308.doi:10.1215/ijm/1256055111.
  6. ^Peter Nylen; Tin-Yau Tam; Frank Uhlig (1993). "On the eigenvalues of principal submatrices of normal, hermitian and symmetric matrices".Linear and Multilinear Algebra.36 (1):69–78.doi:10.1080/03081089308818276.
  7. ^Bebiano N, Furtado S, da Providência J (2011)."On the eigenvalues of principal submatrices of J-normal matrices".Linear Algebra and Its Applications.435 (12):3101–3114.doi:10.1016/j.laa.2011.05.033.
  8. ^Forrester PJ, Zhang J (2021). "Corank-1 projections and the randomised Horn problem".Tunisian Journal of Mathematics.3:55–73.arXiv:1905.05314.doi:10.2140/tunis.2021.3.55.S2CID 153312446.
  9. ^Denton PB, Parke SJ, Tao T, Zhang X (2021). "Eigenvectors from eigenvalues: A survey of a basic identity in linear algebra".Bulletin of the American Mathematical Society.59: 1.arXiv:1908.03795.doi:10.1090/bull/1722.S2CID 213918682.
  10. ^abcPress, William H.; Teukolsky, Saul A.; Vetterling, William T.; Flannery, Brian P. (1992).Numerical Recipes in C (2nd ed.). Cambridge University Press.ISBN 978-0-521-43108-8.
  11. ^Coakley, Ed S. (May 2013), "A fast divide-and-conquer algorithm for computing the spectra of real symmetric tridiagonal matrices.",Applied and Computational Harmonic Analysis,34 (3):379–414,doi:10.1016/j.acha.2012.06.003
  12. ^Neymeyr, K. (2006), "A geometric theory for preconditioned inverse iteration IV: On the fastest convergence cases.",Linear Algebra Appl.,415 (1):114–139,doi:10.1016/j.laa.2005.06.022
  13. ^Li, T. Y.; Zeng, Zhonggang (1992), "Laguerre's Iteration In Solving The Symmetric Tridiagonal Eigenproblem - Revisited",SIAM Journal on Scientific Computing
  14. ^Chu, Moody T. (1988), "A Note on the Homotopy Method for Linear Algebraic Eigenvalue Problems",Linear Algebra Appl.,105:225–236,doi:10.1016/0024-3795(88)90015-8
  15. ^Dhillon, Inderjit S.; Parlett, Beresford N.; Vömel, Christof (2006),"The Design and Implementation of the MRRR Algorithm"(PDF),ACM Transactions on Mathematical Software,32 (4):533–560,doi:10.1145/1186785.1186788,S2CID 2410736
  16. ^Delattre, B.; Barthélemy, Q.; Araujo, A.; Allauzen, A. (2023),"Efficient Bound of Lipschitz Constant for Convolutional Layers by Gram Iteration",Proceedings of the 40th International Conference on Machine Learning:7513–7532
  17. ^Smith, Oliver K. (April 1961), "Eigenvalues of a symmetric 3 × 3 matrix.",Communications of the ACM,4 (4): 168,doi:10.1145/355578.366316,S2CID 37815415

Further reading

[edit]
Key concepts
Problems
Hardware
Software
Retrieved from "https://en.wikipedia.org/w/index.php?title=Eigenvalue_algorithm&oldid=1292279373"
Category:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp