Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Eigendecomposition of a matrix

From Wikipedia, the free encyclopedia
(Redirected fromEigenvalue decomposition)
Matrix decomposition
This articlemay be too technical for most readers to understand. Pleasehelp improve it tomake it understandable to non-experts, without removing the technical details.(August 2024) (Learn how and when to remove this message)

Inlinear algebra,eigendecomposition is thefactorization of amatrix into acanonical form, whereby the matrix is represented in terms of itseigenvalues and eigenvectors. Onlydiagonalizable matrices can be factorized in this way. When the matrix being factorized is anormal or realsymmetric matrix, the decomposition is called "spectral decomposition", derived from thespectral theorem.

Fundamental theory of matrix eigenvectors and eigenvalues

[edit]
See also:Eigenvalue, eigenvector and eigenspace

A (nonzero) vectorv of dimensionN is an eigenvector of a squareN ×N matrixA if it satisfies alinear equation of the formAv=λv{\displaystyle \mathbf {A} \mathbf {v} =\lambda \mathbf {v} }for some scalarλ. Thenλ is called the eigenvalue corresponding tov. Geometrically speaking, the eigenvectors ofA are the vectors thatA merely elongates or shrinks, and the amount that they elongate/shrink by is the eigenvalue. The above equation is called the eigenvalue equation or the eigenvalue problem.

This yields an equation for the eigenvaluesp(λ)=det(AλI)=0.{\displaystyle p\left(\lambda \right)=\det \left(\mathbf {A} -\lambda \mathbf {I} \right)=0.}We callp(λ) thecharacteristic polynomial, and the equation, called the characteristic equation, is anNth-order polynomial equation in the unknownλ. This equation will haveNλ distinct solutions, where1 ≤NλN. The set of solutions, that is, the eigenvalues, is called thespectrum ofA.[1][2][3]

If the field of scalars isalgebraically closed, then we canfactorp asp(λ)=(λλ1)n1(λλ2)n2(λλNλ)nNλ=0.{\displaystyle p(\lambda )=\left(\lambda -\lambda _{1}\right)^{n_{1}}\left(\lambda -\lambda _{2}\right)^{n_{2}}\cdots \left(\lambda -\lambda _{N_{\lambda }}\right)^{n_{N_{\lambda }}}=0.}The integerni is termed thealgebraic multiplicity of eigenvalueλi. The algebraic multiplicities sum toN:i=1Nλni=N.{\textstyle \sum _{i=1}^{N_{\lambda }}{n_{i}}=N.}

For each eigenvalueλi, we have a specific eigenvalue equation(AλiI)v=0.{\displaystyle \left(\mathbf {A} -\lambda _{i}\mathbf {I} \right)\mathbf {v} =0.}There will be1 ≤minilinearly independent solutions to each eigenvalue equation. The linear combinations of themi solutions (except the one which gives the zero vector) are the eigenvectors associated with the eigenvalueλi. The integermi is termed thegeometric multiplicity ofλi. It is important to keep in mind that the algebraic multiplicityni and geometric multiplicitymi may or may not be equal, but we always havemini. The simplest case is of course whenmi =ni = 1. The total number of linearly independent eigenvectors,Nv, can be calculated by summing the geometric multiplicitiesi=1Nλmi=Nv.{\displaystyle \sum _{i=1}^{N_{\lambda }}{m_{i}}=N_{\mathbf {v} }.}

The eigenvectors can be indexed by eigenvalues, using a double index, withvij being thejth eigenvector for theith eigenvalue. The eigenvectors can also be indexed using the simpler notation of a single indexvk, withk = 1, 2, ...,Nv.

Eigendecomposition of a matrix

[edit]

LetA be a squaren ×n matrix withn linearly independent eigenvectorsqi (wherei = 1, ...,n). ThenA can befactored asA=QΛQ1{\displaystyle \mathbf {A} =\mathbf {Q} \mathbf {\Lambda } \mathbf {Q} ^{-1}}whereQ is the squaren ×n matrix whoseith column is the eigenvectorqi ofA, andΛ is thediagonal matrix whose diagonal elements are the corresponding eigenvalues,Λii =λi. Note that onlydiagonalizable matrices can be factorized in this way. For example, thedefective matrix[1101]{\displaystyle \left[{\begin{smallmatrix}1&1\\0&1\end{smallmatrix}}\right]} (which is ashear matrix) cannot be diagonalized.

Then eigenvectorsqi are usually normalized, but they don't have to be. A non-normalized set ofn eigenvectors,vi can also be used as the columns ofQ. That can be understood by noting that the magnitude of the eigenvectors inQ gets canceled in the decomposition by the presence ofQ−1. If one of the eigenvaluesλi has multiple linearly independent eigenvectors (that is, the geometric multiplicity ofλi is greater than 1), then these eigenvectors for this eigenvalueλi can be chosen to be mutuallyorthogonal; however, if two eigenvectors belong to two different eigenvalues, it may be impossible for them to be orthogonal to each other (see Example below). One special case is that ifA is a normal matrix, then by the spectral theorem, it's always possible to diagonalizeA in anorthonormal basis{qi}.

The decomposition can be derived from the fundamental property of eigenvectors:Av=λvAQ=QΛA=QΛQ1.{\displaystyle {\begin{aligned}\mathbf {A} \mathbf {v} &=\lambda \mathbf {v} \\\mathbf {A} \mathbf {Q} &=\mathbf {Q} \mathbf {\Lambda } \\\mathbf {A} &=\mathbf {Q} \mathbf {\Lambda } \mathbf {Q} ^{-1}.\end{aligned}}}The linearly independent eigenvectorsqi with nonzero eigenvalues form a basis (not necessarily orthonormal) for all possible productsAx, forxCn, which is the same as theimage (orrange) of the correspondingmatrix transformation, and also thecolumn space of the matrixA. The number of linearly independent eigenvectorsqi with nonzero eigenvalues is equal to therank of the matrixA, and also the dimension of the image (or range) of the corresponding matrix transformation, as well as its column space.

The linearly independent eigenvectorsqi with an eigenvalue of zero form a basis (which can be chosen to be orthonormal) for thenull space (also known as the kernel) of the matrix transformationA.

Example

[edit]

The 2 × 2 real matrixAA=[1013]{\displaystyle \mathbf {A} ={\begin{bmatrix}1&0\\1&3\\\end{bmatrix}}}may be decomposed into a diagonal matrix through multiplication of a non-singular matrixQQ=[abcd]R2×2.{\displaystyle \mathbf {Q} ={\begin{bmatrix}a&b\\c&d\end{bmatrix}}\in \mathbb {R} ^{2\times 2}.}

Then[abcd]1[1013][abcd]=[x00y],{\displaystyle {\begin{bmatrix}a&b\\c&d\end{bmatrix}}^{-1}{\begin{bmatrix}1&0\\1&3\end{bmatrix}}{\begin{bmatrix}a&b\\c&d\end{bmatrix}}={\begin{bmatrix}x&0\\0&y\end{bmatrix}},}for some real diagonal matrix[x00y]{\displaystyle \left[{\begin{smallmatrix}x&0\\0&y\end{smallmatrix}}\right]}.

Multiplying both sides of the equation on the left byQ:[1013][abcd]=[abcd][x00y].{\displaystyle {\begin{bmatrix}1&0\\1&3\end{bmatrix}}{\begin{bmatrix}a&b\\c&d\end{bmatrix}}={\begin{bmatrix}a&b\\c&d\end{bmatrix}}{\begin{bmatrix}x&0\\0&y\end{bmatrix}}.}The above equation can be decomposed into twosimultaneous equations:{[1013][ac]=[axcx][1013][bd]=[bydy].{\displaystyle {\begin{cases}{\begin{bmatrix}1&0\\1&3\end{bmatrix}}{\begin{bmatrix}a\\c\end{bmatrix}}={\begin{bmatrix}ax\\cx\end{bmatrix}}\\[1.2ex]{\begin{bmatrix}1&0\\1&3\end{bmatrix}}{\begin{bmatrix}b\\d\end{bmatrix}}={\begin{bmatrix}by\\dy\end{bmatrix}}\end{cases}}.}Factoring out theeigenvaluesx andy:{[1013][ac]=x[ac][1013][bd]=y[bd]{\displaystyle {\begin{cases}{\begin{bmatrix}1&0\\1&3\end{bmatrix}}{\begin{bmatrix}a\\c\end{bmatrix}}=x{\begin{bmatrix}a\\c\end{bmatrix}}\\[1.2ex]{\begin{bmatrix}1&0\\1&3\end{bmatrix}}{\begin{bmatrix}b\\d\end{bmatrix}}=y{\begin{bmatrix}b\\d\end{bmatrix}}\end{cases}}}Lettinga=[ac],b=[bd],{\displaystyle \mathbf {a} ={\begin{bmatrix}a\\c\end{bmatrix}},\quad \mathbf {b} ={\begin{bmatrix}b\\d\end{bmatrix}},}this gives us two vector equations:{Aa=xaAb=yb{\displaystyle {\begin{cases}\mathbf {A} \mathbf {a} =x\mathbf {a} \\\mathbf {A} \mathbf {b} =y\mathbf {b} \end{cases}}}And can be represented by a single vector equation involving two solutions as eigenvalues:Au=λu{\displaystyle \mathbf {A} \mathbf {u} =\lambda \mathbf {u} }whereλ represents the two eigenvaluesx andy, andu represents the vectorsa andb.

Shiftingλu to the left hand side and factoringu out(AλI)u=0{\displaystyle \left(\mathbf {A} -\lambda \mathbf {I} \right)\mathbf {u} =\mathbf {0} }SinceQ is non-singular, it is essential thatu is nonzero. Therefore,det(AλI)=0{\displaystyle \det(\mathbf {A} -\lambda \mathbf {I} )=0}Thus(1λ)(3λ)=0{\displaystyle (1-\lambda )(3-\lambda )=0}giving us the solutions of the eigenvalues for the matrixA asλ = 1 orλ = 3, and the resulting diagonal matrix from the eigendecomposition ofA is thus[1003]{\displaystyle \left[{\begin{smallmatrix}1&0\\0&3\end{smallmatrix}}\right]}.

Putting the solutions back into the above simultaneous equations{[1013][ac]=1[ac][1013][bd]=3[bd]{\displaystyle {\begin{cases}{\begin{bmatrix}1&0\\1&3\end{bmatrix}}{\begin{bmatrix}a\\c\end{bmatrix}}=1{\begin{bmatrix}a\\c\end{bmatrix}}\\[1.2ex]{\begin{bmatrix}1&0\\1&3\end{bmatrix}}{\begin{bmatrix}b\\d\end{bmatrix}}=3{\begin{bmatrix}b\\d\end{bmatrix}}\end{cases}}}

Solving the equations, we havea=2candb=0,c,dR{0}.{\displaystyle a=-2c\quad {\text{and}}\quad b=0,\qquad c,d\in \mathbb {R} \setminus \{0\}.}Thus the matrixQ required for the eigendecomposition ofA isQ=[2c0cd],c,dR{0},{\displaystyle \mathbf {Q} ={\begin{bmatrix}-2c&0\\c&d\end{bmatrix}},\qquad c,d\in \mathbb {R} \setminus \{0\},}that is:[2c0cd]1[1013][2c0cd]=[1003],c,dR{0}.{\displaystyle {\begin{bmatrix}-2c&0\\c&d\end{bmatrix}}^{-1}{\begin{bmatrix}1&0\\1&3\end{bmatrix}}{\begin{bmatrix}-2c&0\\c&d\end{bmatrix}}={\begin{bmatrix}1&0\\0&3\end{bmatrix}},\qquad c,d\in \mathbb {R} \setminus \{0\}.}The exclusion of the number 0 from the set of real numbers,R{\displaystyle \mathbb {R} }, is necessary to ensure that the matrixQ{\displaystyle \mathbf {Q} } is non-singular.

Matrix inverse via eigendecomposition

[edit]
Main article:Inverse matrix

If a matrixA can be eigendecomposed and if none of its eigenvalues are zero, thenA isinvertible and its inverse is given byA1=QΛ1Q1{\displaystyle \mathbf {A} ^{-1}=\mathbf {Q} \mathbf {\Lambda } ^{-1}\mathbf {Q} ^{-1}}IfA{\displaystyle \mathbf {A} } is a symmetric matrix, sinceQ{\displaystyle \mathbf {Q} } is formed from the eigenvectors ofA{\displaystyle \mathbf {A} },Q{\displaystyle \mathbf {Q} } is guaranteed to be anorthogonal matrix, thereforeQ1=QT{\displaystyle \mathbf {Q} ^{-1}=\mathbf {Q} ^{\mathrm {T} }}. Furthermore, becauseΛ is adiagonal matrix, its inverse is easy to calculate:[Λ1]ii=1λi{\displaystyle \left[\mathbf {\Lambda } ^{-1}\right]_{ii}={\frac {1}{\lambda _{i}}}}

Practical implications

[edit]

When eigendecomposition is used on a matrix of measured, realdata, theinverse may be less valid when all eigenvalues are used unmodified in the form above. This is because as eigenvalues become relatively small, their contribution to the inversion is large. Those near zero or at the "noise" of the measurement system will have undue influence and could hamper solutions (detection) using the inverse.[4]

Two mitigations have been proposed: truncating small or zero eigenvalues, and extending the lowest reliable eigenvalue to those below it. See alsoTikhonov regularization as a statistically motivated but biased method for rolling off eigenvalues as they become dominated by noise.

The first mitigation method is similar to a sparse sample of the original matrix, removing components that are not considered valuable. However, if the solution or detection process is near the noise level, truncating may remove components that influence the desired solution.

The second mitigation extends the eigenvalue so that lower values have much less influence over inversion, but do still contribute, such that solutions near the noise will still be found.

The reliable eigenvalue can be found by assuming that eigenvalues of extremely similar and low value are a good representation of measurement noise (which is assumed low for most systems).

If the eigenvalues are rank-sorted by value, then the reliable eigenvalue can be found by minimization of theLaplacian of the sorted eigenvalues:[5]min|2λs|{\displaystyle \min \left|\nabla ^{2}\lambda _{\mathrm {s} }\right|}where the eigenvalues are subscripted with ans to denote being sorted. The position of the minimization is the lowest reliable eigenvalue. In measurement systems, the square root of this reliable eigenvalue is the average noise over the components of the system.

Functional calculus

[edit]

The eigendecomposition allows for much easier computation ofpower series of matrices. Iff (x) is given byf(x)=a0+a1x+a2x2+{\displaystyle f(x)=a_{0}+a_{1}x+a_{2}x^{2}+\cdots }then we know thatf(A)=Qf(Λ)Q1{\displaystyle f\!\left(\mathbf {A} \right)=\mathbf {Q} \,f\!\left(\mathbf {\Lambda } \right)\mathbf {Q} ^{-1}}BecauseΛ is adiagonal matrix, functions ofΛ are very easy to calculate:[f(Λ)]ii=f(λi){\displaystyle \left[f\left(\mathbf {\Lambda } \right)\right]_{ii}=f\left(\lambda _{i}\right)}

The off-diagonal elements off (Λ) are zero; that is,f (Λ) is also a diagonal matrix. Therefore, calculatingf (A) reduces to just calculating the function on each of the eigenvalues.

A similar technique works more generally with theholomorphic functional calculus, usingA1=QΛ1Q1{\displaystyle \mathbf {A} ^{-1}=\mathbf {Q} \mathbf {\Lambda } ^{-1}\mathbf {Q} ^{-1}}fromabove. Once again, we find that[f(Λ)]ii=f(λi){\displaystyle \left[f\left(\mathbf {\Lambda } \right)\right]_{ii}=f\left(\lambda _{i}\right)}

Examples

[edit]

A2=(QΛQ1)(QΛQ1)=QΛ(Q1Q)ΛQ1=QΛ2Q1An=QΛnQ1expA=Qexp(Λ)Q1{\displaystyle {\begin{aligned}\mathbf {A} ^{2}&=\left(\mathbf {Q} \mathbf {\Lambda } \mathbf {Q} ^{-1}\right)\left(\mathbf {Q} \mathbf {\Lambda } \mathbf {Q} ^{-1}\right)=\mathbf {Q} \mathbf {\Lambda } \left(\mathbf {Q} ^{-1}\mathbf {Q} \right)\mathbf {\Lambda } \mathbf {Q} ^{-1}=\mathbf {Q} \mathbf {\Lambda } ^{2}\mathbf {Q} ^{-1}\\[1.2ex]\mathbf {A} ^{n}&=\mathbf {Q} \mathbf {\Lambda } ^{n}\mathbf {Q} ^{-1}\\[1.2ex]\exp \mathbf {A} &=\mathbf {Q} \exp(\mathbf {\Lambda } )\mathbf {Q} ^{-1}\end{aligned}}}which are examples for the functionsf(x)=x2,f(x)=xn,f(x)=expx{\displaystyle f(x)=x^{2},\;f(x)=x^{n},\;f(x)=\exp {x}}. Furthermore,expA{\displaystyle \exp {\mathbf {A} }} is thematrix exponential.

Decomposition for spectral matrices

[edit]
Main article:Spectral theorem
[icon]
This sectionneeds expansion. You can help byadding missing information.(June 2008)

Spectral matrices are matrices that possess distinct eigenvalues and a complete set of eigenvectors. This characteristic allows spectral matrices to be fully diagonalizable, meaning they can be decomposed into simpler forms using eigendecomposition. This decomposition process reveals fundamental insights into the matrix's structure and behavior, particularly in fields such as quantum mechanics, signal processing, and numerical analysis.[6]

Normal matrices

[edit]

A complex-valued square matrixA{\displaystyle A} isnormal (meaning ,AA=AA{\displaystyle \mathbf {A} ^{*}\mathbf {A} =\mathbf {A} \mathbf {A} ^{*}}, whereA{\displaystyle \mathbf {A} ^{*}} is theconjugate transpose) if and only if it can be decomposed asA=UΛU{\displaystyle \mathbf {A} =\mathbf {U} \mathbf {\Lambda } \mathbf {U} ^{*}}, whereU{\displaystyle \mathbf {U} } is aunitary matrix (meaningU=U1{\displaystyle \mathbf {U} ^{*}=\mathbf {U} ^{-1}}) andΛ={\displaystyle \mathbf {\Lambda } =} diag(λ1,,λn{\displaystyle \lambda _{1},\ldots ,\lambda _{n}}) is adiagonal matrix.[7] The columnsu1,,un{\displaystyle \mathbf {u} _{1},\cdots ,\mathbf {u} _{n}} ofU{\displaystyle \mathbf {U} } form anorthonormal basis and are eigenvectors ofA{\displaystyle \mathbf {A} } with corresponding eigenvaluesλ1,,λn{\displaystyle \lambda _{1},\ldots ,\lambda _{n}}.[8]

For example, consider the 2 x 2 normal matrixA=[1221]{\displaystyle \mathbf {A} ={\begin{bmatrix}1&2\\2&1\end{bmatrix}}}.

The eigenvalues areλ1=3{\displaystyle \lambda _{1}=3} andλ2=1{\displaystyle \lambda _{2}=-1}.

The (normalized) eigenvectors corresponding to these eigenvalues areu1=12[11]{\displaystyle \mathbf {u} _{1}={\frac {1}{\sqrt {2}}}{\begin{bmatrix}1\\1\end{bmatrix}}} andu2=12[11]{\displaystyle \mathbf {u} _{2}={\frac {1}{\sqrt {2}}}{\begin{bmatrix}-1\\1\end{bmatrix}}}.

The diagonalization isA=UΛU{\displaystyle \mathbf {A} =\mathbf {U} \mathbf {\Lambda } \mathbf {U} ^{*}}, whereU=[1/21/21/21/2]{\displaystyle \mathbf {U} ={\begin{bmatrix}1/{\sqrt {2}}&1/{\sqrt {2}}\\1/{\sqrt {2}}&-1/{\sqrt {2}}\end{bmatrix}}},Λ={\displaystyle \mathbf {\Lambda } =}[3001]{\displaystyle {\begin{bmatrix}3&0\\0&-1\end{bmatrix}}} andU=U1={\displaystyle \mathbf {U} ^{*}=\mathbf {U} ^{-1}=}[1/21/21/21/2]{\displaystyle {\begin{bmatrix}1/{\sqrt {2}}&1/{\sqrt {2}}\\1/{\sqrt {2}}&-1/{\sqrt {2}}\end{bmatrix}}}.

The verification isUΛU={\displaystyle \mathbf {U} \mathbf {\Lambda } \mathbf {U} ^{*}=}[1/21/21/21/2]{\displaystyle {\begin{bmatrix}1/{\sqrt {2}}&1/{\sqrt {2}}\\1/{\sqrt {2}}&-1/{\sqrt {2}}\end{bmatrix}}}[3001]{\displaystyle {\begin{bmatrix}3&0\\0&-1\end{bmatrix}}}[1/21/21/21/2]{\displaystyle {\begin{bmatrix}1/{\sqrt {2}}&1/{\sqrt {2}}\\1/{\sqrt {2}}&-1/{\sqrt {2}}\end{bmatrix}}}=[1221]=A{\displaystyle ={\begin{bmatrix}1&2\\2&1\end{bmatrix}}=\mathbf {A} }.

This example illustrates the process of diagonalizing a normal matrixA{\displaystyle \mathbf {A} } by finding its eigenvalues and eigenvectors, forming the unitary matrixU{\displaystyle \mathbf {U} }, the diagonal matrixΛ{\displaystyle \mathbf {\Lambda } }, and verifying the decomposition.

Subsets of important classes of matrices

Real symmetric matrices

[edit]

As a special case, for everyn ×n realsymmetric matrix, the eigenvalues are real and the eigenvectors can be chosen real andorthonormal. Thus a real symmetric matrixA can be decomposed asA=QΛQT{\displaystyle \mathbf {A} =\mathbf {Q} \mathbf {\Lambda } \mathbf {Q} ^{\mathsf {T}}}, whereQ is anorthogonal matrix whose columns are the real, orthonormal eigenvectors ofA, andΛ is a diagonal matrix whose entries are the eigenvalues ofA.[9]

Diagonalizable matrices

[edit]

Diagonalizable matrices can be decomposed using eigendecomposition, provided they have a full set of linearly independent eigenvectors. They can be expressed asA=PDP1{\displaystyle \mathbf {A} =\mathbf {P} \mathbf {D} \mathbf {P} ^{-1}}, whereP{\displaystyle \mathbf {P} } is a matrix whose columns are eigenvectors ofA{\displaystyle \mathbf {A} }, andD{\displaystyle \mathbf {D} } is a diagonal matrix consisting of the corresponding eigenvalues ofA{\displaystyle \mathbf {A} }.[8]

Positive definite matrices

[edit]

Positivedefinite matrices are matrices for which all eigenvalues are positive. They can be decomposed asA=LLT{\displaystyle \mathbf {A} =\mathbf {L} \mathbf {L} ^{\mathsf {T}}} using theCholesky decomposition, whereL{\displaystyle \mathbf {L} } is a lower triangular matrix.[10]

Unitary and Hermitian matrices

[edit]

Unitary matrices satisfyUU=I{\displaystyle \mathbf {U} \mathbf {U} ^{*}=\mathbf {I} } (real case) orUU=I{\displaystyle \mathbf {U} \mathbf {U} ^{\dagger }=\mathbf {I} } (complex case), whereU{\displaystyle \mathbf {U} ^{*}}denotes theconjugate transpose andU{\displaystyle \mathbf {U} ^{\dagger }}denotes the conjugate transpose. They diagonalize usingunitary transformations.[8]

Hermitian matrices satisfyH=H{\displaystyle \mathbf {H} =\mathbf {H} ^{\dagger }}, whereH{\displaystyle \mathbf {H} ^{\dagger }}denotes the conjugate transpose. They can be diagonalized using unitary ororthogonal matrices.[8]

Useful facts

[edit]

Useful facts regarding eigenvalues

[edit]

Useful facts regarding eigenvectors

[edit]
  • IfA isHermitian and full-rank, the basis of eigenvectors may be chosen to be mutuallyorthogonal. The eigenvalues are real.
  • The eigenvectors ofA−1 are the same as the eigenvectors ofA.
  • Eigenvectors are only defined up to a multiplicative constant. That is, ifAv =λv thencv is also an eigenvector for any scalarc ≠ 0. In particular,v andev (for anyθ) are also eigenvectors.
  • In the case of degenerate eigenvalues (an eigenvalue having more than one eigenvector), the eigenvectors have an additional freedom of linear transformation, that is to say, any linear (orthonormal) combination of eigenvectors sharing an eigenvalue (in the degenerate subspace) is itself an eigenvector (in the subspace).

Useful facts regarding eigendecomposition

[edit]
  • A can be eigendecomposed if and only if the number of linearly independent eigenvectors,Nv, equals the dimension of an eigenvector:Nv =N
  • If the field of scalars is algebraically closed and ifp(λ) has no repeated roots, that is, ifNλ=N,{\displaystyle N_{\lambda }=N,} thenA can be eigendecomposed.
  • The statement "A can be eigendecomposed" doesnot imply thatA has an inverse as some eigenvalues may be zero, makingA not invertible.
  • The statement "A has an inverse" doesnot imply thatA can be eigendecomposed. A counterexample is[1101]{\displaystyle \left[{\begin{smallmatrix}1&1\\0&1\end{smallmatrix}}\right]}, which is an invertibledefective matrix.

Useful facts regarding matrix inverse

[edit]

Numerical computations

[edit]
Further information:Eigenvalue algorithm

Numerical computation of eigenvalues

[edit]

Suppose that we want to compute the eigenvalues of a given matrix. If the matrix is small, we can compute them symbolically using thecharacteristic polynomial. However, this is often impossible for larger matrices, in which case we must use anumerical method.

In practice, eigenvalues of large matrices are not computed using the characteristic polynomial. Computing the polynomial becomes expensive in itself, and exact (symbolic) roots of a high-degree polynomial can be difficult to compute and express: theAbel–Ruffini theorem implies that the roots of high-degree (5 or above) polynomials cannot in general be expressed simply usingnth roots. Therefore, general algorithms to find eigenvectors and eigenvalues areiterative.

Iterative numerical algorithms for approximating roots of polynomials exist, such asNewton's method, but in general it is impractical to compute the characteristic polynomial and then apply these methods. One reason is that smallround-off errors in the coefficients of the characteristic polynomial can lead to large errors in the eigenvalues and eigenvectors: the roots are an extremelyill-conditioned function of the coefficients.[11]

A simple and accurate iterative method is thepower method: arandom vectorv is chosen and a sequence ofunit vectors is computed asAvAv,A2vA2v,A3vA3v,{\displaystyle {\frac {\mathbf {A} \mathbf {v} }{\left\|\mathbf {A} \mathbf {v} \right\|}},{\frac {\mathbf {A} ^{2}\mathbf {v} }{\left\|\mathbf {A} ^{2}\mathbf {v} \right\|}},{\frac {\mathbf {A} ^{3}\mathbf {v} }{\left\|\mathbf {A} ^{3}\mathbf {v} \right\|}},\ldots }

Thissequence willalmost always converge to an eigenvector corresponding to the eigenvalue of greatest magnitude, provided thatv has a nonzero component of this eigenvector in the eigenvector basis (and also provided that there is only one eigenvalue of greatest magnitude). This simple algorithm is useful in some practical applications; for example,Google uses it to calculate thepage rank of documents in their search engine.[12] Also, the power method is the starting point for many more sophisticated algorithms. For instance, by keeping not just the last vector in the sequence, but instead looking at thespan ofall the vectors in the sequence, one can get a better (faster converging) approximation for the eigenvector, and this idea is the basis ofArnoldi iteration.[11] Alternatively, the importantQR algorithm is also based on a subtle transformation of a power method.[11]

Numerical computation of eigenvectors

[edit]

Once the eigenvalues are computed, the eigenvectors could be calculated by solving the equation(AλiI)vi,j=0{\displaystyle \left(\mathbf {A} -\lambda _{i}\mathbf {I} \right)\mathbf {v} _{i,j}=\mathbf {0} }usingGaussian elimination orany other method for solvingmatrix equations.

However, in practical large-scale eigenvalue methods, the eigenvectors are usually computed in other ways, as a byproduct of the eigenvalue computation. Inpower iteration, for example, the eigenvector is actually computed before the eigenvalue (which is typically computed by theRayleigh quotient of the eigenvector).[11] In the QR algorithm for aHermitian matrix (or any normal matrix), the orthonormal eigenvectors are obtained as a product of theQ matrices from the steps in the algorithm.[11] (For more general matrices, the QR algorithm yields theSchur decomposition first, from which the eigenvectors can be obtained by abacksubstitution procedure.[13]) For Hermitian matrices, theDivide-and-conquer eigenvalue algorithm is more efficient than the QR algorithm if both eigenvectors and eigenvalues are desired.[11]

Additional topics

[edit]

Generalized eigenspaces

[edit]

Recall that thegeometric multiplicity of an eigenvalue can be described as the dimension of the associated eigenspace, thenullspace ofλIA. The algebraic multiplicity can also be thought of as a dimension: it is the dimension of the associatedgeneralized eigenspace (1st sense), which is the nullspace of the matrix(λIA)k forany sufficiently largek. That is, it is the space ofgeneralized eigenvectors (first sense), where a generalized eigenvector is any vector whicheventually becomes 0 ifλIA is applied to it enough times successively. Any eigenvector is a generalized eigenvector, and so each eigenspace is contained in the associated generalized eigenspace. This provides an easy proof that the geometric multiplicity is always less than or equal to the algebraic multiplicity.

This usage should not be confused with thegeneralized eigenvalue problem described below.

Conjugate eigenvector

[edit]

Aconjugate eigenvector orconeigenvector is a vector sent after transformation to a scalar multiple of its conjugate, where the scalar is called theconjugate eigenvalue orconeigenvalue of the linear transformation. The coneigenvectors and coneigenvalues represent essentially the same information and meaning as the regular eigenvectors and eigenvalues, but arise when an alternative coordinate system is used. The corresponding equation isAv=λv.{\displaystyle \mathbf {A} \mathbf {v} =\lambda \mathbf {v} ^{*}.}For example, in coherent electromagnetic scattering theory, the linear transformationA represents the action performed by the scattering object, and the eigenvectors represent polarization states of the electromagnetic wave. Inoptics, the coordinate system is defined from the wave's viewpoint, known as theForward Scattering Alignment (FSA), and gives rise to a regular eigenvalue equation, whereas inradar, the coordinate system is defined from the radar's viewpoint, known as theBack Scattering Alignment (BSA), and gives rise to a coneigenvalue equation.

Generalized eigenvalue problem

[edit]

Ageneralized eigenvalue problem (second sense) is the problem of finding a (nonzero) vectorv that obeysAv=λBv{\displaystyle \mathbf {A} \mathbf {v} =\lambda \mathbf {B} \mathbf {v} }whereA andB are matrices. Ifv obeys this equation, with someλ, then we callv thegeneralized eigenvector ofA andB (in the second sense), andλ is called thegeneralized eigenvalue ofA andB (in the second sense) which corresponds to the generalized eigenvectorv. The possible values ofλ must obey the following equationdet(AλB)=0.{\displaystyle \det(\mathbf {A} -\lambda \mathbf {B} )=0.}

Ifn linearly independent vectors{v1, …,vn} can be found, such that for everyi ∈ {1, …,n},Avi =λiBvi, then we define the matricesP andD such thatP=[||v1vn||][(v1)1(vn)1(v1)n(vn)n]{\displaystyle P={\begin{bmatrix}|&&|\\\mathbf {v} _{1}&\cdots &\mathbf {v} _{n}\\|&&|\end{bmatrix}}\equiv {\begin{bmatrix}(\mathbf {v} _{1})_{1}&\cdots &(\mathbf {v} _{n})_{1}\\\vdots &&\vdots \\(\mathbf {v} _{1})_{n}&\cdots &(\mathbf {v} _{n})_{n}\end{bmatrix}}}(D)ij={λi,if i=j0,otherwise{\displaystyle (D)_{ij}={\begin{cases}\lambda _{i},&{\text{if }}i=j\\0,&{\text{otherwise}}\end{cases}}}Then the following equality holdsA=BPDP1{\displaystyle \mathbf {A} =\mathbf {B} \mathbf {P} \mathbf {D} \mathbf {P} ^{-1}}And the proof isAP=A[||v1vn||]=[||Av1Avn||]=[||λ1Bv1λnBvn||]=[||Bv1Bvn||]D=BPD{\displaystyle \mathbf {A} \mathbf {P} =\mathbf {A} {\begin{bmatrix}|&&|\\\mathbf {v} _{1}&\cdots &\mathbf {v} _{n}\\|&&|\end{bmatrix}}={\begin{bmatrix}|&&|\\A\mathbf {v} _{1}&\cdots &A\mathbf {v} _{n}\\|&&|\end{bmatrix}}={\begin{bmatrix}|&&|\\\lambda _{1}B\mathbf {v} _{1}&\cdots &\lambda _{n}B\mathbf {v} _{n}\\|&&|\end{bmatrix}}={\begin{bmatrix}|&&|\\B\mathbf {v} _{1}&\cdots &B\mathbf {v} _{n}\\|&&|\end{bmatrix}}\mathbf {D} =\mathbf {B} \mathbf {P} \mathbf {D} }

And sinceP is invertible, we multiply the equation from the right by its inverse, finishing the proof.

The set of matrices of the formAλB, whereλ is a complex number, is called apencil; the termmatrix pencil can also refer to the pair(A,B) of matrices.[14]

IfB is invertible, then the original problem can be written in the formB1Av=λv{\displaystyle \mathbf {B} ^{-1}\mathbf {A} \mathbf {v} =\lambda \mathbf {v} }which is a standard eigenvalue problem. However, in most situations it is preferable not to perform the inversion, but rather to solve the generalized eigenvalue problem as stated originally. This is especially important ifA andB areHermitian matrices, since in this caseB−1A is not generally Hermitian and important properties of the solution are no longer apparent.

IfA andB are both symmetric or Hermitian, andB is also apositive-definite matrix, the eigenvaluesλi are real and eigenvectorsv1 andv2 with distinct eigenvalues areB-orthogonal (v1*Bv2 = 0).[15] In this case, eigenvectors can be chosen so that the matrixP defined above satisfiesPBP=I{\displaystyle \mathbf {P} ^{*}\mathbf {B} \mathbf {P} =\mathbf {I} } orPPB=I,{\displaystyle \mathbf {P} \mathbf {P} ^{*}\mathbf {B} =\mathbf {I} ,}and there exists abasis of generalized eigenvectors (it is not adefective problem).[14] This case is sometimes called aHermitian definite pencil ordefinite pencil.[14]

See also

[edit]

Notes

[edit]
  1. ^Golub, Gene H.;Van Loan, Charles F. (1996),Matrix Computations (3rd ed.), Baltimore:Johns Hopkins University Press, p. 310,ISBN 978-0-8018-5414-9
  2. ^Kreyszig, Erwin (1972),Advanced Engineering Mathematics (3rd ed.), New York:Wiley, p. 273,ISBN 978-0-471-50728-4
  3. ^Nering, Evar D. (1970).Linear Algebra and Matrix Theory (2nd ed.). New York:Wiley. p. 270.LCCN 76091646.
  4. ^Hayde, A. F.; Twede, D. R. (2002). Shen, Sylvia S. (ed.). "Observations on relationship between eigenvalues, instrument noise and detection performance".Imaging Spectrometry VIII. Proceedings of SPIE.4816: 355.Bibcode:2002SPIE.4816..355H.doi:10.1117/12.453777.S2CID 120953647.
  5. ^Twede, D. R.; Hayden, A. F. (2004). Shen, Sylvia S; Lewis, Paul E (eds.). "Refinement and generalization of the extension method of covariance matrix inversion by regularization".Imaging Spectrometry IX. Proceedings of SPIE.5159: 299.Bibcode:2004SPIE.5159..299T.doi:10.1117/12.506993.S2CID 123123072.
  6. ^Allaire, Grégoire (2008).Numerical linear algebra. Springer.ISBN 978-0-387-34159-0.
  7. ^Horn & Johnson 1985, p. 133, Theorem 2.5.3
  8. ^abcdShores, Thomas S (2006)."Applied linear algebra and matrix analysis".
  9. ^Horn & Johnson 1985, p. 136, Corollary 2.5.11
  10. ^Carl D. Meyer (2023).Matrix analysis and applied linear algebra (2nd ed.). Society for Industrial and Applied Mathematics.ISBN 9781611977431.
  11. ^abcdefTrefethen, Lloyd N.; Bau, David (1997).Numerical Linear Algebra. SIAM.ISBN 978-0-89871-361-9.
  12. ^Ipsen, Ilse, and Rebecca M. Wills,Analysis and Computation of Google's PageRankArchived 2018-09-21 at theWayback Machine, 7th IMACS International Symposium on Iterative Methods in Scientific Computing, Fields Institute, Toronto, Canada, 5–8 May 2005.
  13. ^Quarteroni, Alfio; Sacco, Riccardo; Saleri, Fausto (2000). "section 5.8.2".Numerical Mathematics. Springer. p. 15.ISBN 978-0-387-98959-4.
  14. ^abcBai, Z.;Demmel, J.;Dongarra, J.; Ruhe, A.; Van Der Vorst, H., eds. (2000). "Generalized Hermitian Eigenvalue Problems".Templates for the Solution of Algebraic Eigenvalue Problems: A Practical Guide. Philadelphia: SIAM.ISBN 978-0-89871-471-5. Archived fromthe original on 2010-08-21. Retrieved2022-09-09.
  15. ^Parlett, Beresford N. (1998).The symmetric eigenvalue problem (Reprint. ed.). Philadelphia: Society for Industrial and Applied Mathematics. p. 345.doi:10.1137/1.9781611971163.ISBN 978-0-89871-402-9.

References

[edit]

External links

[edit]
Retrieved from "https://en.wikipedia.org/w/index.php?title=Eigendecomposition_of_a_matrix&oldid=1323812043"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp