Inlinear algebra, asquare matrix is calleddiagonalizable ornon-defective if it issimilar to adiagonal matrix. That is, if there exists aninvertible matrix and a diagonal matrix such that. This is equivalent to. (Such, are not unique.) This property exists for any linear map: for afinite-dimensionalvector space, alinear map is calleddiagonalizable if there exists anordered basis of consisting ofeigenvectors of. These definitions are equivalent: if has amatrix representation as above, then the column vectors of form a basis consisting of eigenvectors of, and the diagonal entries of are the correspondingeigenvalues of; with respect to this eigenvector basis, is represented by.
Diagonalization is the process of finding the above and and makes many subsequent computations easier. One can raise a diagonal matrix to a power by simply raising the diagonal entries to that power. Thedeterminant of a diagonal matrix is simply the product of all diagonal entries. Such computations generalize easily to.
The geometric transformation represented by a diagonalizable matrix is aninhomogeneous dilation (oranisotropic scaling). That is, it canscale the space by a different amount in different directions. The direction of each eigenvector is scaled by a factor given by the corresponding eigenvalue.
A square matrix that is not diagonalizable is calleddefective. It can happen that a matrix withreal entries is defective over the real numbers, meaning that is impossible for any invertible and diagonal with real entries, but it is possible withcomplex entries, so that is diagonalizable over the complex numbers. For example, this is the case for a genericrotation matrix.
Many results for diagonalizable matrices hold only over analgebraically closed field (such as the complex numbers). In this case, diagonalizable matrices aredense in the space of all matrices, which means any defective matrix can be deformed into a diagonalizable matrix by a smallperturbation; and theJordan–Chevalley decomposition states that any matrix is uniquely the sum of a diagonalizable matrix and anilpotent matrix. Over an algebraically closed field, diagonalizable matrices are equivalent tosemi-simple matrices.
A square matrix with entries in afield is calleddiagonalizable ornondefective if there exists an invertible matrix (i.e. an element of thegeneral linear group GLn(F)),, such that is a diagonal matrix.
The fundamental fact about diagonalizable maps and matrices is expressed by the following:
The following sufficient (but not necessary) condition is often useful.
Let be a matrix over. If is diagonalizable, then so is any power of it. Conversely, if is invertible, is algebraically closed, and is diagonalizable for some that is not an integer multiple of the characteristic of, then is diagonalizable. Proof: If is diagonalizable, then is annihilated by some polynomial, which has no multiple root (since) and is divided by the minimal polynomial of.
Over the complex numbers, almost every matrix is diagonalizable. More precisely: the set of complex matrices that arenot diagonalizable over, considered as asubset of, hasLebesgue measure zero. One can also say that the diagonalizable matrices form a dense subset with respect to theZariski topology: the non-diagonalizable matrices lie inside thevanishing set of thediscriminant of the characteristic polynomial, which is ahypersurface. From that follows also density in the usual (strong) topology given by anorm. The same is not true over.
TheJordan–Chevalley decomposition expresses an operator as the sum of its semisimple (i.e., diagonalizable) part and itsnilpotent part. Hence, a matrix is diagonalizable if and only if its nilpotent part is zero. Put in another way, a matrix is diagonalizable if each block in itsJordan form has no nilpotent part; i.e., each "block" is a one-by-one matrix.
Consider the two following arbitrary bases and. Suppose that there exists a linear transformation represented by a matrix which is written with respect to basis E. Suppose also that there exists the following eigen-equation:
The alpha eigenvectors are written also with respect to the E basis. Since the set F is both a set of eigenvectors for matrix A and it spans some arbitrary vector space, then we say that there exists a matrix which is a diagonal matrix that is similar to. In other words, is a diagonalizable matrix if the matrix is written in the basis F. We perform the change of basis calculation using the transition matrix, which changes basis from E to F as follows:
,
where is the transition matrix from E-basis to F-basis. The inverse can then be equated to a new transition matrix which changes basis from F to E instead and so we have the following relationship :
Both and transition matrices are invertible. Thus we can manipulate the matrices in the following fashion:The matrix will be denoted as, which is still in the E-basis. Similarly, the diagonal matrix is in the F-basis.

If a matrix can be diagonalized, that is,
then:
The transition matrix S has the E-basis vectors as columns written in the basis F. Inversely, the inverse transition matrix P has F-basis vectors written in the basis of E so that we can represent P in block matrix form in the following manner:
as a result we can write:
In block matrix form, we can consider the A-matrix to be a matrix of 1x1 dimensions whilst P is a 1xn dimensional matrix. The D-matrix can be written in full form with all the diagonal elements as an nxn dimensional matrix:
Performing the above matrix multiplication we end up with the following result:Taking each component of the block matrix individually on both sides, we end up with the following:
So the column vectors of areright eigenvectors of, and the corresponding diagonal entry is the correspondingeigenvalue. The invertibility of also suggests that the eigenvectors arelinearly independent and form a basis of. This is the necessary and sufficient condition for diagonalizability and the canonical approach of diagonalization. Therow vectors of are theleft eigenvectors of.
When a complex matrix is aHermitian matrix (or more generally anormal matrix), eigenvectors of can be chosen to form anorthonormal basis of, and can be chosen to be aunitary matrix. If in addition, is a realsymmetric matrix, then its eigenvectors can be chosen to be an orthonormal basis of and can be chosen to be anorthogonal matrix.
For most practical work matrices are diagonalized numerically using computer software.Many algorithms exist to accomplish this.
A set of matrices is said to besimultaneously diagonalizable if there exists a single invertible matrix such that is a diagonal matrix for every in the set. The following theorem characterizes simultaneously diagonalizable matrices: A set of diagonalizablematrices commutes if and only if the set is simultaneously diagonalizable.[1]: p. 64
The set of all diagonalizable matrices (over) with is not simultaneously diagonalizable. For instance, the matrices
are diagonalizable but not simultaneously diagonalizable because they do not commute.
A set consists of commutingnormal matrices if and only if it is simultaneously diagonalizable by aunitary matrix; that is, there exists a unitary matrix such that is diagonal for every in the set.
In the language ofLie theory, a set of simultaneously diagonalizable matrices generates atoral Lie algebra.
In general, arotation matrix is not diagonalizable over the reals, but allrotation matrices are diagonalizable over the complex field. Even if a matrix is not diagonalizable, it is always possible to "do the best one can", and find a matrix with the same properties consisting of eigenvalues on the leading diagonal, and either ones or zeroes on the superdiagonal – known asJordan normal form.
Some matrices are not diagonalizable over any field, most notably nonzeronilpotent matrices. This happens more generally if thealgebraic and geometric multiplicities of an eigenvalue do not coincide. For instance, consider
This matrix is not diagonalizable: there is no matrix such that is a diagonal matrix. Indeed, has one eigenvalue (namely zero) and this eigenvalue has algebraic multiplicity 2 and geometric multiplicity 1.
Some real matrices are not diagonalizable over the reals. Consider for instance the matrix
The matrix does not have any real eigenvalues, so there is noreal matrix such that is a diagonal matrix. However, we can diagonalize if we allow complex numbers. Indeed, if we take
then is diagonal. It is easy to find that is the rotation matrix which rotates counterclockwise by angle
Note that the above examples show that the sum of diagonalizable matrices need not be diagonalizable.
Diagonalizing a matrix is the same process as finding itseigenvalues and eigenvectors, in the case that the eigenvectors form a basis. For example, consider the matrix
The roots of thecharacteristic polynomial are the eigenvalues. Solving the linear system gives the eigenvectors and, while gives; that is, for. These vectors form a basis of, so we can assemble them as the column vectors of achange-of-basis matrix to get:We may see this equation in terms of transformations: takes the standard basis to the eigenbasis,, so we have:so that has the standard basis as its eigenvectors, which is the defining property of.
Note that there is no preferred order of the eigenvectors in; changing the order of theeigenvectors in just changes the order of theeigenvalues in the diagonalized form of.[2]
Diagonalization can be used to efficiently compute the powers of a matrix:
and the latter is easy to calculate since it only involves the powers of a diagonal matrix. For example, for the matrix with eigenvalues in the example above we compute:
This approach can be generalized tomatrix exponential and othermatrix functions that can be defined as power series. For example, defining, we have:
This is particularly useful in finding closed form expressions for terms oflinear recursive sequences, such as theFibonacci numbers.
For example, consider the following matrix:
Calculating the various powers of reveals a surprising pattern:
The above phenomenon can be explained by diagonalizing. To accomplish this, we need a basis of consisting of eigenvectors of. One such eigenvector basis is given by
whereei denotes the standard basis ofRn. The reverse change of basis is given by
Straightforward calculations show that
Thus,a andb are the eigenvalues corresponding tou andv, respectively. By linearity of matrix multiplication, we have that
Switching back to the standard basis, we have
The preceding relations, expressed in matrix form, are
thereby explaining the above phenomenon.
Inquantum mechanical andquantum chemical computations matrix diagonalization is one of the most frequently applied numerical processes. The basic reason is that the time-independentSchrödinger equation is an eigenvalue equation, albeit in most of the physical situations on an infinite dimensionalHilbert space.
A very common approximation is to truncate (or project) the Hilbert space to finite dimension, after which the Schrödinger equation can be formulated as an eigenvalue problem of a real symmetric, or complex Hermitian matrix. Formally this approximation is founded on thevariational principle, valid for Hamiltonians that are bounded from below.
First-order perturbation theory also leads to matrix eigenvalue problem for degenerate states.
Matrices can be generalized tolinear operators. A diagonal matrix can be generalized to diagonal operators on Hilbert spaces.
Let be a Hilbert space. An operator is a diagonal operator iff there exists an orthonormal basis of, such that for some.
For any, define thep-Schatten norm as follows. Let be an operator, then, where is thetrace. The p-Schatten class is the set of all operators with finite p-Schatten norm.
Weyl,[3]von Neumann,[4] and Kuroda,[5] showed the following:
For any, any self-adjoint operator on a Hilbert space, and any, there exists a diagonal operator, such that.
In other words, any self-adjoint operator is an infinitesimal perturbation from a diagonal operator, where "infinitesimal" is in the sense of p-Schatten norm. In particular, since the Hilbert–Schmidt operator class is the 2-Schatten class, this means that any self-adjoint operator is diagonalizable after a perturbation by an infinitesimal Hilbert–Schmidt operator.In fact, the above result could be further generalized:
For anynorm ideal that is not the trace class, with norm, any self-adjoint operator on a Hilbert space, and any, there exists a diagonal operator, such that.
The result is false for (thetrace class). This is a simple corollary of the Kato[6]–Rosenblum[7][8]: Theorem XI.8 theorem, which states that if is self-adjoint, and is trace class, then have the sameabsolutely continuous part of the spectrum. The result is sharp, however, in the sense that if has no absolutely continuous part, then itcan be diagonalized after perturbation by an infinitesimal trace class operator.[9]
Forsimultaneous diagonalization, it's known that, given a finite list of self-adjoint operators that commute with each other, for any, there exists a sequence of diagonal operators, such that, where is the n-Schatten norm. Note that[10]