Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Diagonalizable matrix

From Wikipedia, the free encyclopedia
Matrices similar to diagonal matrices
This article is about matrix diagonalization in linear algebra. For other uses, seeDiagonalization.

Inlinear algebra, asquare matrixA{\displaystyle A} is calleddiagonalizable ornon-defective if it issimilar to adiagonal matrix. That is, if there exists aninvertible matrixP{\displaystyle P} and a diagonal matrixD{\displaystyle D} such thatP1AP=D{\displaystyle P^{-1}AP=D}. This is equivalent toA=PDP1{\displaystyle A=PDP^{-1}}. (SuchP{\displaystyle P},D{\displaystyle D} are not unique.) This property exists for any linear map: for afinite-dimensionalvector spaceV{\displaystyle V}, alinear mapT:VV{\displaystyle T:V\to V} is calleddiagonalizable if there exists anordered basis ofV{\displaystyle V} consisting ofeigenvectors ofT{\displaystyle T}. These definitions are equivalent: ifT{\displaystyle T} has amatrix representationA=PDP1{\displaystyle A=PDP^{-1}} as above, then the column vectors ofP{\displaystyle P} form a basis consisting of eigenvectors ofT{\displaystyle T}, and the diagonal entries ofD{\displaystyle D} are the correspondingeigenvalues ofT{\displaystyle T}; with respect to this eigenvector basis,T{\displaystyle T} is represented byD{\displaystyle D}.

Diagonalization is the process of finding the aboveP{\displaystyle P} andD{\displaystyle D} and makes many subsequent computations easier. One can raise a diagonal matrixD{\displaystyle D} to a power by simply raising the diagonal entries to that power. Thedeterminant of a diagonal matrix is simply the product of all diagonal entries. Such computations generalize easily toA=PDP1{\displaystyle A=PDP^{-1}}.

The geometric transformation represented by a diagonalizable matrix is aninhomogeneous dilation (oranisotropic scaling). That is, it canscale the space by a different amount in different directions. The direction of each eigenvector is scaled by a factor given by the corresponding eigenvalue.

A square matrix that is not diagonalizable is calleddefective. It can happen that a matrixA{\displaystyle A} withreal entries is defective over the real numbers, meaning thatA=PDP1{\displaystyle A=PDP^{-1}} is impossible for any invertibleP{\displaystyle P} and diagonalD{\displaystyle D} with real entries, but it is possible withcomplex entries, so thatA{\displaystyle A} is diagonalizable over the complex numbers. For example, this is the case for a genericrotation matrix.

Many results for diagonalizable matrices hold only over analgebraically closed field (such as the complex numbers). In this case, diagonalizable matrices aredense in the space of all matrices, which means any defective matrix can be deformed into a diagonalizable matrix by a smallperturbation; and theJordan–Chevalley decomposition states that any matrix is uniquely the sum of a diagonalizable matrix and anilpotent matrix. Over an algebraically closed field, diagonalizable matrices are equivalent tosemi-simple matrices.

Definition

[edit]

A squaren×n{\displaystyle n\times n} matrixA{\displaystyle A} with entries in afieldF{\displaystyle F} is calleddiagonalizable ornondefective if there exists ann×n{\displaystyle n\times n} invertible matrix (i.e. an element of thegeneral linear group GLn(F)),P{\displaystyle P}, such thatP1AP{\displaystyle P^{-1}AP} is a diagonal matrix.

Characterization

[edit]

The fundamental fact about diagonalizable maps and matrices is expressed by the following:

The following sufficient (but not necessary) condition is often useful.

LetA{\displaystyle A} be a matrix overF{\displaystyle F}. IfA{\displaystyle A} is diagonalizable, then so is any power of it. Conversely, ifA{\displaystyle A} is invertible,F{\displaystyle F} is algebraically closed, andAn{\displaystyle A^{n}} is diagonalizable for somen{\displaystyle n} that is not an integer multiple of the characteristic ofF{\displaystyle F}, thenA{\displaystyle A} is diagonalizable. Proof: IfAn{\displaystyle A^{n}} is diagonalizable, thenA{\displaystyle A} is annihilated by some polynomial(xnλ1)(xnλk){\displaystyle \left(x^{n}-\lambda _{1}\right)\cdots \left(x^{n}-\lambda _{k}\right)}, which has no multiple root (sinceλj0{\displaystyle \lambda _{j}\neq 0}) and is divided by the minimal polynomial ofA{\displaystyle A}.

Over the complex numbersC{\displaystyle \mathbb {C} }, almost every matrix is diagonalizable. More precisely: the set of complexn×n{\displaystyle n\times n} matrices that arenot diagonalizable overC{\displaystyle \mathbb {C} }, considered as asubset ofCn×n{\displaystyle \mathbb {C} ^{n\times n}}, hasLebesgue measure zero. One can also say that the diagonalizable matrices form a dense subset with respect to theZariski topology: the non-diagonalizable matrices lie inside thevanishing set of thediscriminant of the characteristic polynomial, which is ahypersurface. From that follows also density in the usual (strong) topology given by anorm. The same is not true overR{\displaystyle \mathbb {R} }.

TheJordan–Chevalley decomposition expresses an operator as the sum of its semisimple (i.e., diagonalizable) part and itsnilpotent part. Hence, a matrix is diagonalizable if and only if its nilpotent part is zero. Put in another way, a matrix is diagonalizable if each block in itsJordan form has no nilpotent part; i.e., each "block" is a one-by-one matrix.

Diagonalization

[edit]
See also:Eigendecomposition of a matrix

Consider the two following arbitrary basesE={ei|i[n]}{\displaystyle E=\{{{\boldsymbol {e}}_{i}|\forall i\in [n]}\}} andF={αi|i[n]}{\displaystyle F=\{{{\boldsymbol {\alpha }}_{i}|\forall i\in [n]}\}}. Suppose that there exists a linear transformation represented by a matrixAE{\displaystyle A_{E}} which is written with respect to basis E. Suppose also that there exists the following eigen-equation:

AEαE,i=λiαE,i{\displaystyle A_{E}{\boldsymbol {\alpha }}_{E,i}=\lambda _{i}{\boldsymbol {\alpha }}_{E,i}}

The alpha eigenvectors are written also with respect to the E basis. Since the set F is both a set of eigenvectors for matrix A and it spans some arbitrary vector space, then we say that there exists a matrixDF{\displaystyle D_{F}} which is a diagonal matrix that is similar toAE{\displaystyle A_{E}}. In other words,AE{\displaystyle A_{E}} is a diagonalizable matrix if the matrix is written in the basis F. We perform the change of basis calculation using the transition matrixS{\displaystyle S}, which changes basis from E to F as follows:

DF=SEF AE SE1F{\displaystyle D_{F}=S_{E}^{F}\ A_{E}\ S_{E}^{-1F}},

whereSEF{\displaystyle S_{E}^{F}} is the transition matrix from E-basis to F-basis. The inverse can then be equated to a new transition matrixP{\displaystyle P} which changes basis from F to E instead and so we have the following relationship :

SE1F=PFE{\displaystyle S_{E}^{-1F}=P_{F}^{E}}

BothS{\displaystyle S} andP{\displaystyle P} transition matrices are invertible. Thus we can manipulate the matrices in the following fashion:D=S AE S1D=P1 AE P{\displaystyle {\begin{aligned}D=S\ A_{E}\ S^{-1}\\D=P^{-1}\ A_{E}\ P\end{aligned}}}The matrixAE{\displaystyle A_{E}} will be denoted asA{\displaystyle A}, which is still in the E-basis. Similarly, the diagonal matrix is in the F-basis.

The diagonalization of a symmetric matrix can be interpreted as a rotation of the axes to align them with the eigenvectors.

If a matrixA{\displaystyle A} can be diagonalized, that is,

P1AP=[λ1000λ2000λn]=D,{\displaystyle P^{-1}AP={\begin{bmatrix}\lambda _{1}&0&\cdots &0\\0&\lambda _{2}&\cdots &0\\\vdots &\vdots &\ddots &\vdots \\0&0&\cdots &\lambda _{n}\end{bmatrix}}=D,}

then:

AP=P[λ1000λ2000λn].{\displaystyle AP=P{\begin{bmatrix}\lambda _{1}&0&\cdots &0\\0&\lambda _{2}&\cdots &0\\\vdots &\vdots &\ddots &\vdots \\0&0&\cdots &\lambda _{n}\end{bmatrix}}.}

The transition matrix S has the E-basis vectors as columns written in the basis F. Inversely, the inverse transition matrix P has F-basis vectorsαi{\displaystyle {\boldsymbol {\alpha }}_{i}} written in the basis of E so that we can represent P in block matrix form in the following manner:

P=[αE,1αE,2αE,n],{\displaystyle P={\begin{bmatrix}{\boldsymbol {\alpha }}_{E,1}&{\boldsymbol {\alpha }}_{E,2}&\cdots &{\boldsymbol {\alpha }}_{E,n}\end{bmatrix}},}

as a result we can write:A[αE,1αE,2αE,n]=[αE,1αE,2αE,n]D.{\displaystyle {\begin{aligned}A{\begin{bmatrix}{\boldsymbol {\alpha }}_{E,1}&{\boldsymbol {\alpha }}_{E,2}&\cdots &{\boldsymbol {\alpha }}_{E,n}\end{bmatrix}}={\begin{bmatrix}{\boldsymbol {\alpha }}_{E,1}&{\boldsymbol {\alpha }}_{E,2}&\cdots &{\boldsymbol {\alpha }}_{E,n}\end{bmatrix}}D.\end{aligned}}}

In block matrix form, we can consider the A-matrix to be a matrix of 1x1 dimensions whilst P is a 1xn dimensional matrix. The D-matrix can be written in full form with all the diagonal elements as an nxn dimensional matrix:

A[αE,1αE,2αE,n]=[αE,1αE,2αE,n][λ1000λ2000λn].{\displaystyle A{\begin{bmatrix}{\boldsymbol {\alpha }}_{E,1}&{\boldsymbol {\alpha }}_{E,2}&\cdots &{\boldsymbol {\alpha }}_{E,n}\end{bmatrix}}={\begin{bmatrix}{\boldsymbol {\alpha }}_{E,1}&{\boldsymbol {\alpha }}_{E,2}&\cdots &{\boldsymbol {\alpha }}_{E,n}\end{bmatrix}}{\begin{bmatrix}\lambda _{1}&0&\cdots &0\\0&\lambda _{2}&\cdots &0\\\vdots &\vdots &\ddots &\vdots \\0&0&\cdots &\lambda _{n}\end{bmatrix}}.}

Performing the above matrix multiplication we end up with the following result:A[α1α2αn]=[λ1α1λ2α2λnαn]{\displaystyle {\begin{aligned}A{\begin{bmatrix}{\boldsymbol {\alpha }}_{1}&{\boldsymbol {\alpha }}_{2}&\cdots &{\boldsymbol {\alpha }}_{n}\end{bmatrix}}={\begin{bmatrix}\lambda _{1}{\boldsymbol {\alpha }}_{1}&\lambda _{2}{\boldsymbol {\alpha }}_{2}&\cdots &\lambda _{n}{\boldsymbol {\alpha }}_{n}\end{bmatrix}}\end{aligned}}}Taking each component of the block matrix individually on both sides, we end up with the following:

Aαi=λiαi(i=1,2,,n).{\displaystyle A{\boldsymbol {\alpha }}_{i}=\lambda _{i}{\boldsymbol {\alpha }}_{i}\qquad (i=1,2,\dots ,n).}

So the column vectors ofP{\displaystyle P} areright eigenvectors ofA{\displaystyle A}, and the corresponding diagonal entry is the correspondingeigenvalue. The invertibility ofP{\displaystyle P} also suggests that the eigenvectors arelinearly independent and form a basis ofFn{\displaystyle F^{n}}. This is the necessary and sufficient condition for diagonalizability and the canonical approach of diagonalization. Therow vectors ofP1{\displaystyle P^{-1}} are theleft eigenvectors ofA{\displaystyle A}.

When a complex matrixACn×n{\displaystyle A\in \mathbb {C} ^{n\times n}} is aHermitian matrix (or more generally anormal matrix), eigenvectors ofA{\displaystyle A} can be chosen to form anorthonormal basis ofCn{\displaystyle \mathbb {C} ^{n}}, andP{\displaystyle P} can be chosen to be aunitary matrix. If in addition,ARn×n{\displaystyle A\in \mathbb {R} ^{n\times n}} is a realsymmetric matrix, then its eigenvectors can be chosen to be an orthonormal basis ofRn{\displaystyle \mathbb {R} ^{n}} andP{\displaystyle P} can be chosen to be anorthogonal matrix.

For most practical work matrices are diagonalized numerically using computer software.Many algorithms exist to accomplish this.

Simultaneous diagonalization

[edit]
See also:Simultaneous triangularisability,Weight (representation theory), andPositive definite matrix

A set of matrices is said to besimultaneously diagonalizable if there exists a single invertible matrixP{\displaystyle P} such thatP1AP{\displaystyle P^{-1}AP} is a diagonal matrix for everyA{\displaystyle A} in the set. The following theorem characterizes simultaneously diagonalizable matrices: A set of diagonalizablematrices commutes if and only if the set is simultaneously diagonalizable.[1]: p. 64 

The set of alln×n{\displaystyle n\times n} diagonalizable matrices (overC{\displaystyle \mathbb {C} }) withn>1{\displaystyle n>1} is not simultaneously diagonalizable. For instance, the matrices

[1000]and[1100]{\displaystyle {\begin{bmatrix}1&0\\0&0\end{bmatrix}}\quad {\text{and}}\quad {\begin{bmatrix}1&1\\0&0\end{bmatrix}}}

are diagonalizable but not simultaneously diagonalizable because they do not commute.

A set consists of commutingnormal matrices if and only if it is simultaneously diagonalizable by aunitary matrix; that is, there exists a unitary matrixU{\displaystyle U} such thatUAU{\displaystyle U^{*}AU} is diagonal for everyA{\displaystyle A} in the set.

In the language ofLie theory, a set of simultaneously diagonalizable matrices generates atoral Lie algebra.

Examples

[edit]

Diagonalizable matrices

[edit]

Matrices that are not diagonalizable

[edit]

In general, arotation matrix is not diagonalizable over the reals, but allrotation matrices are diagonalizable over the complex field. Even if a matrix is not diagonalizable, it is always possible to "do the best one can", and find a matrix with the same properties consisting of eigenvalues on the leading diagonal, and either ones or zeroes on the superdiagonal – known asJordan normal form.

Some matrices are not diagonalizable over any field, most notably nonzeronilpotent matrices. This happens more generally if thealgebraic and geometric multiplicities of an eigenvalue do not coincide. For instance, consider

C=[0100].{\displaystyle C={\begin{bmatrix}0&1\\0&0\end{bmatrix}}.}

This matrix is not diagonalizable: there is no matrixU{\displaystyle U} such thatU1CU{\displaystyle U^{-1}CU} is a diagonal matrix. Indeed,C{\displaystyle C} has one eigenvalue (namely zero) and this eigenvalue has algebraic multiplicity 2 and geometric multiplicity 1.

Some real matrices are not diagonalizable over the reals. Consider for instance the matrix

B=[0110].{\displaystyle B=\left[{\begin{array}{rr}0&1\\\!-1&0\end{array}}\right].}

The matrixB{\displaystyle B} does not have any real eigenvalues, so there is noreal matrixQ{\displaystyle Q} such thatQ1BQ{\displaystyle Q^{-1}BQ} is a diagonal matrix. However, we can diagonalizeB{\displaystyle B} if we allow complex numbers. Indeed, if we take

Q=[1ii1],{\displaystyle Q={\begin{bmatrix}1&i\\i&1\end{bmatrix}},}

thenQ1BQ{\displaystyle Q^{-1}BQ} is diagonal. It is easy to find thatB{\displaystyle B} is the rotation matrix which rotates counterclockwise by angleθ=π2{\textstyle \theta =-{\frac {\pi }{2}}}

Note that the above examples show that the sum of diagonalizable matrices need not be diagonalizable.

How to diagonalize a matrix

[edit]

Diagonalizing a matrix is the same process as finding itseigenvalues and eigenvectors, in the case that the eigenvectors form a basis. For example, consider the matrix

A=[012010113].{\displaystyle A=\left[{\begin{array}{rrr}0&1&\!\!\!-2\\0&1&0\\1&\!\!\!-1&3\end{array}}\right].}

The roots of thecharacteristic polynomialp(λ)=det(λIA){\displaystyle p(\lambda )=\det(\lambda I-A)} are the eigenvaluesλ1=1,λ2=1,λ3=2{\displaystyle \lambda _{1}=1,\lambda _{2}=1,\lambda _{3}=2}. Solving the linear system(1IA)v=0{\displaystyle \left(1I-A\right)\mathbf {v} =\mathbf {0} } gives the eigenvectorsv1=(1,1,0){\displaystyle \mathbf {v} _{1}=(1,1,0)} andv2=(0,2,1){\displaystyle \mathbf {v} _{2}=(0,2,1)}, while(2IA)v=0{\displaystyle \left(2I-A\right)\mathbf {v} =\mathbf {0} } givesv3=(1,0,1){\displaystyle \mathbf {v} _{3}=(1,0,-1)}; that is,Avi=λivi{\displaystyle A\mathbf {v} _{i}=\lambda _{i}\mathbf {v} _{i}} fori=1,2,3{\displaystyle i=1,2,3}. These vectors form a basis ofV=R3{\displaystyle V=\mathbb {R} ^{3}}, so we can assemble them as the column vectors of achange-of-basis matrixP{\displaystyle P} to get:P1AP=[101120011]1[012010113][101120011]=[100010002]=D.{\displaystyle P^{-1}AP=\left[{\begin{array}{rrr}1&0&1\\1&2&0\\0&1&\!\!\!\!-1\end{array}}\right]^{-1}\left[{\begin{array}{rrr}0&1&\!\!\!-2\\0&1&0\\1&\!\!\!-1&3\end{array}}\right]\left[{\begin{array}{rrr}1&\,0&1\\1&2&0\\0&1&\!\!\!\!-1\end{array}}\right]={\begin{bmatrix}1&0&0\\0&1&0\\0&0&2\end{bmatrix}}=D.}We may see this equation in terms of transformations:P{\displaystyle P} takes the standard basis to the eigenbasis,Pei=vi{\displaystyle P\mathbf {e} _{i}=\mathbf {v} _{i}}, so we have:P1APei=P1Avi=P1(λivi)=λiei,{\displaystyle P^{-1}AP\mathbf {e} _{i}=P^{-1}A\mathbf {v} _{i}=P^{-1}(\lambda _{i}\mathbf {v} _{i})=\lambda _{i}\mathbf {e} _{i},}so thatP1AP{\displaystyle P^{-1}AP} has the standard basis as its eigenvectors, which is the defining property ofD{\displaystyle D}.

Note that there is no preferred order of the eigenvectors inP{\displaystyle P}; changing the order of theeigenvectors inP{\displaystyle P} just changes the order of theeigenvalues in the diagonalized form ofA{\displaystyle A}.[2]

Application to matrix functions

[edit]

Diagonalization can be used to efficiently compute the powers of a matrixA=PDP1{\displaystyle A=PDP^{-1}}:

Ak=(PDP1)k=(PDP1)(PDP1)(PDP1)=PD(P1P)D(P1P)(P1P)DP1=PDkP1,{\displaystyle {\begin{aligned}A^{k}&=\left(PDP^{-1}\right)^{k}=\left(PDP^{-1}\right)\left(PDP^{-1}\right)\cdots \left(PDP^{-1}\right)\\&=PD\left(P^{-1}P\right)D\left(P^{-1}P\right)\cdots \left(P^{-1}P\right)DP^{-1}=PD^{k}P^{-1},\end{aligned}}}

and the latter is easy to calculate since it only involves the powers of a diagonal matrix. For example, for the matrixA{\displaystyle A} with eigenvaluesλ=1,1,2{\displaystyle \lambda =1,1,2} in the example above we compute:

Ak=PDkP1=[101120011][1k0001k0002k][101120011]1=[22k1+2k22k+10101+2k12k1+2k+1].{\displaystyle {\begin{aligned}A^{k}=PD^{k}P^{-1}&=\left[{\begin{array}{rrr}1&\,0&1\\1&2&0\\0&1&\!\!\!\!-1\end{array}}\right]{\begin{bmatrix}1^{k}&0&0\\0&1^{k}&0\\0&0&2^{k}\end{bmatrix}}\left[{\begin{array}{rrr}1&\,0&1\\1&2&0\\0&1&\!\!\!\!-1\end{array}}\right]^{-1}\\[1em]&={\begin{bmatrix}2-2^{k}&-1+2^{k}&2-2^{k+1}\\0&1&0\\-1+2^{k}&1-2^{k}&-1+2^{k+1}\end{bmatrix}}.\end{aligned}}}

This approach can be generalized tomatrix exponential and othermatrix functions that can be defined as power series. For example, definingexp(A)=I+A+12!A2+13!A3+{\textstyle \exp(A)=I+A+{\frac {1}{2!}}A^{2}+{\frac {1}{3!}}A^{3}+\cdots }, we have:

exp(A)=Pexp(D)P1=[101120011][e1000e1000e2][101120011]1=[2ee2e+e22e2e20e0e+e2ee2e+2e2].{\displaystyle {\begin{aligned}\exp(A)=P\exp(D)P^{-1}&=\left[{\begin{array}{rrr}1&\,0&1\\1&2&0\\0&1&\!\!\!\!-1\end{array}}\right]{\begin{bmatrix}e^{1}&0&0\\0&e^{1}&0\\0&0&e^{2}\end{bmatrix}}\left[{\begin{array}{rrr}1&\,0&1\\1&2&0\\0&1&\!\!\!\!-1\end{array}}\right]^{-1}\\[1em]&={\begin{bmatrix}2e-e^{2}&-e+e^{2}&2e-2e^{2}\\0&e&0\\-e+e^{2}&e-e^{2}&-e+2e^{2}\end{bmatrix}}.\end{aligned}}}

This is particularly useful in finding closed form expressions for terms oflinear recursive sequences, such as theFibonacci numbers.

Particular application

[edit]

For example, consider the following matrix:

M=[aba0b].{\displaystyle M={\begin{bmatrix}a&b-a\\0&b\end{bmatrix}}.}

Calculating the various powers ofM{\displaystyle M} reveals a surprising pattern:

M2=[a2b2a20b2],M3=[a3b3a30b3],M4=[a4b4a40b4],{\displaystyle M^{2}={\begin{bmatrix}a^{2}&b^{2}-a^{2}\\0&b^{2}\end{bmatrix}},\quad M^{3}={\begin{bmatrix}a^{3}&b^{3}-a^{3}\\0&b^{3}\end{bmatrix}},\quad M^{4}={\begin{bmatrix}a^{4}&b^{4}-a^{4}\\0&b^{4}\end{bmatrix}},\quad \ldots }

The above phenomenon can be explained by diagonalizingM{\displaystyle M}. To accomplish this, we need a basis ofR2{\displaystyle \mathbb {R} ^{2}} consisting of eigenvectors ofM{\displaystyle M}. One such eigenvector basis is given by

u=[10]=e1,v=[11]=e1+e2,{\displaystyle \mathbf {u} ={\begin{bmatrix}1\\0\end{bmatrix}}=\mathbf {e} _{1},\quad \mathbf {v} ={\begin{bmatrix}1\\1\end{bmatrix}}=\mathbf {e} _{1}+\mathbf {e} _{2},}

whereei denotes the standard basis ofRn. The reverse change of basis is given by

e1=u,e2=vu.{\displaystyle \mathbf {e} _{1}=\mathbf {u} ,\qquad \mathbf {e} _{2}=\mathbf {v} -\mathbf {u} .}

Straightforward calculations show that

Mu=au,Mv=bv.{\displaystyle M\mathbf {u} =a\mathbf {u} ,\qquad M\mathbf {v} =b\mathbf {v} .}

Thus,a andb are the eigenvalues corresponding tou andv, respectively. By linearity of matrix multiplication, we have that

Mnu=anu,Mnv=bnv.{\displaystyle M^{n}\mathbf {u} =a^{n}\mathbf {u} ,\qquad M^{n}\mathbf {v} =b^{n}\mathbf {v} .}

Switching back to the standard basis, we have

Mne1=Mnu=ane1,Mne2=Mn(vu)=bnvanu=(bnan)e1+bne2.{\displaystyle {\begin{aligned}M^{n}\mathbf {e} _{1}&=M^{n}\mathbf {u} =a^{n}\mathbf {e} _{1},\\M^{n}\mathbf {e} _{2}&=M^{n}\left(\mathbf {v} -\mathbf {u} \right)=b^{n}\mathbf {v} -a^{n}\mathbf {u} =\left(b^{n}-a^{n}\right)\mathbf {e} _{1}+b^{n}\mathbf {e} _{2}.\end{aligned}}}

The preceding relations, expressed in matrix form, are

Mn=[anbnan0bn],{\displaystyle M^{n}={\begin{bmatrix}a^{n}&b^{n}-a^{n}\\0&b^{n}\end{bmatrix}},}

thereby explaining the above phenomenon.

Quantum mechanical application

[edit]

Inquantum mechanical andquantum chemical computations matrix diagonalization is one of the most frequently applied numerical processes. The basic reason is that the time-independentSchrödinger equation is an eigenvalue equation, albeit in most of the physical situations on an infinite dimensionalHilbert space.

A very common approximation is to truncate (or project) the Hilbert space to finite dimension, after which the Schrödinger equation can be formulated as an eigenvalue problem of a real symmetric, or complex Hermitian matrix. Formally this approximation is founded on thevariational principle, valid for Hamiltonians that are bounded from below.

First-order perturbation theory also leads to matrix eigenvalue problem for degenerate states.

Operator theory

[edit]

Matrices can be generalized tolinear operators. A diagonal matrix can be generalized to diagonal operators on Hilbert spaces.

LetH{\displaystyle H} be a Hilbert space. An operatorD:HH{\displaystyle D:H\to H} is a diagonal operator iff there exists an orthonormal basis(en)n{\displaystyle (e_{n})_{n}} ofH{\displaystyle H}, such thatDen=λnen{\displaystyle De_{n}=\lambda _{n}e_{n}} for someλnC{\displaystyle \lambda _{n}\in \mathbb {C} }.

For anyp1{\displaystyle p\geq 1}, define thep-Schatten norm as follows. LetT:HH{\displaystyle T:H\to H} be an operator, thenTp:=Tr(|T|p)1/p{\displaystyle \|T\|_{p}:=\operatorname {Tr} (|T|^{p})^{1/p}}, whereTr{\displaystyle \operatorname {Tr} } is thetrace. The p-Schatten class is the set of all operators with finite p-Schatten norm.

Weyl,[3]von Neumann,[4] and Kuroda,[5] showed the following:

For anyp>1{\displaystyle p>1}, any self-adjoint operatorT{\displaystyle T} on a Hilbert spaceH{\displaystyle H}, and anyϵ>0{\displaystyle \epsilon >0}, there exists a diagonal operatorD{\displaystyle D}, such thatTDpϵ{\displaystyle \|T-D\|_{p}\leq \epsilon }.

In other words, any self-adjoint operator is an infinitesimal perturbation from a diagonal operator, where "infinitesimal" is in the sense of p-Schatten norm. In particular, since the Hilbert–Schmidt operator class is the 2-Schatten class, this means that any self-adjoint operator is diagonalizable after a perturbation by an infinitesimal Hilbert–Schmidt operator.In fact, the above result could be further generalized:

For anynorm ideal that is not the trace class, with normJ{\displaystyle \|\cdot \|_{J}}, any self-adjoint operatorT{\displaystyle T} on a Hilbert spaceH{\displaystyle H}, and anyϵ>0{\displaystyle \epsilon >0}, there exists a diagonal operatorD{\displaystyle D}, such thatTDJϵ{\displaystyle \|T-D\|_{J}\leq \epsilon }.

The result is false forp=1{\displaystyle p=1} (thetrace class). This is a simple corollary of the Kato[6]–Rosenblum[7][8]: Theorem XI.8  theorem, which states that ifT{\displaystyle T} is self-adjoint, andA{\displaystyle A} is trace class, thenT,T+A{\displaystyle T,T+A} have the sameabsolutely continuous part of the spectrum. The result is sharp, however, in the sense that ifT{\displaystyle T} has no absolutely continuous part, then itcan be diagonalized after perturbation by an infinitesimal trace class operator.[9]

Forsimultaneous diagonalization, it's known that, given a finite list ofT1,,Tn{\displaystyle T_{1},\dots ,T_{n}} self-adjoint operators that commute with each other, for anyϵ>0{\displaystyle \epsilon >0}, there exists a sequence of diagonal operatorsD1,,Dn{\displaystyle D_{1},\dots ,D_{n}}, such thatT1D1nϵ,,TnDnnϵ{\displaystyle \|T_{1}-D_{1}\|_{n}\leq \epsilon ,\dots ,\|T_{n}-D_{n}\|_{n}\leq \epsilon }, wheren{\displaystyle \|\cdot \|_{n}} is the n-Schatten norm. Note thatn2{\displaystyle n\geq 2}[10]

See also

[edit]

References

[edit]
  1. ^Horn, Roger A.; Johnson, Charles R. (2013).Matrix Analysis, second edition. Cambridge University Press.ISBN 9780521839402.
  2. ^Anton, H.; Rorres, C. (22 Feb 2000).Elementary Linear Algebra (Applications Version) (8th ed.). John Wiley & Sons.ISBN 978-0-471-17052-5.
  3. ^Von Weyl, Hermann (December 1909)."Über beschränkte quadratische formen, deren differenz vollstetig ist".Rendiconti del Circolo Matematico di Palermo (in German).27 (1):373–392.doi:10.1007/BF03019655.ISSN 0009-725X.
  4. ^von Neumann, John (1935). "Charakterisierung des Spektrums eines Integraloperators" [Characterization of the spectrum of an integral operator].Actualités Scientifiques et Industrielles (in German).229:3–20.
  5. ^Kuroda, Shige Toshi (1958-01-01)."On a theorem of Weyl-von Neumann".Proceedings of the Japan Academy, Series A, Mathematical Sciences.34 (1).doi:10.3792/pja/1195524841.ISSN 0386-2194.
  6. ^Kato, Tosio (1957)."Perturbation of Continuous Spectra by Trace Class Operators".Proceedings of the Japan Academy.33 (5):260–264.doi:10.3792/pja/1195525063.
  7. ^Rosenblum, Marvin (1957)."Perturbation of the continuous spectrum and unitary equivalence".Pacific J. Math.7 (4):997–1010.doi:10.2140/pjm.1957.7.997.
  8. ^Reed, Michael;Simon, Barry (May 12, 1979).Scattering Theory. Methods of Modern Mathematical Physics. Vol. 3 (1st ed.). Academic Press.ISBN 978-0125850032.
  9. ^Carey, R. W.; Pincus, J. D. (1976)."Unitary Equivalence Modulo the Trace Class for Self-Adjoint Operators".American Journal of Mathematics.98 (2):481–514.doi:10.2307/2373898.ISSN 0002-9327.JSTOR 2373898.
  10. ^Voiculescu, Dan (1990-06-01)."On the existence of quasicentral approximate units relative to normed ideals. Part I".Journal of Functional Analysis.91 (1):1–36.doi:10.1016/0022-1236(90)90047-O.ISSN 0022-1236.
Matrix classes
Explicitly constrained entries
Constant
Conditions oneigenvalues or eigenvectors
Satisfying conditions onproducts orinverses
With specific applications
Used instatistics
Used ingraph theory
Used in science and engineering
Related terms
Retrieved from "https://en.wikipedia.org/w/index.php?title=Diagonalizable_matrix&oldid=1322433364"
Category:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp