Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Characteristic polynomial

From Wikipedia, the free encyclopedia

Polynomial whose roots are the eigenvalues of a matrix
This article is about the characteristic polynomial of a matrix or of an endomorphism of vector spaces. For the characteristic polynomial of a matroid, seeMatroid. For that of a graded poset, seeGraded poset.

Inlinear algebra, thecharacteristic polynomial of asquare matrix is apolynomial which isinvariant undermatrix similarity and has theeigenvalues asroots. It has thedeterminant and thetrace of the matrix among its coefficients. Thecharacteristic polynomial of anendomorphism of a finite-dimensionalvector space is the characteristic polynomial of the matrix of that endomorphism over any basis (that is, the characteristic polynomial does not depend on the choice of abasis). Thecharacteristic equation, also known as thedeterminantal equation,[1][2][3] is the equation obtained by equating the characteristic polynomial to zero.

Inspectral graph theory, thecharacteristic polynomial of agraph is the characteristic polynomial of itsadjacency matrix.[4]

Motivation

[edit]

Eigenvalues and eigenvectors play a fundamental role inlinear algebra, since, given alinear transformation, an eigenvector is a vector whose direction is not changed by the transformation, and the corresponding eigenvalue is the measure of the resulting change of magnitude of the vector.

More precisely, suppose the transformation is represented by a square matrixA.{\displaystyle A.} Then an eigenvectorv{\displaystyle \mathbf {v} } and the corresponding eigenvalueλ{\displaystyle \lambda } must satisfy the equationAv=λv,{\displaystyle A\mathbf {v} =\lambda \mathbf {v} ,}or, equivalently (sinceλv=λIv{\displaystyle \lambda \mathbf {v} =\lambda I\mathbf {v} }),(λIA)v=0{\displaystyle (\lambda I-A)\mathbf {v} =\mathbf {0} }whereI{\displaystyle I} is theidentity matrix, andv0{\displaystyle \mathbf {v} \neq \mathbf {0} }(although the zero vector satisfies this equation for everyλ,{\displaystyle \lambda ,} it is not considered an eigenvector).

It follows that the matrix(λIA){\displaystyle (\lambda I-A)} must besingular, and its determinantdet(λIA)=0{\displaystyle \det(\lambda I-A)=0}must be zero.

In other words, the eigenvalues ofA are theroots ofdet(xIA),{\displaystyle \det(xI-A),}which is amonic polynomial inx of degreen ifA is an×n matrix. This polynomial is thecharacteristic polynomial ofA.

Formal definition

[edit]

Consider ann×n{\displaystyle n\times n} matrixA.{\displaystyle A.} The characteristic polynomial ofA,{\displaystyle A,} denoted bypA(t),{\displaystyle p_{A}(t),} is the polynomial defined by[5]pA(t)=det(tIA){\displaystyle p_{A}(t)=\det(tI-A)}whereI{\displaystyle I} denotes then×n{\displaystyle n\times n}identity matrix.

Some authors define the characteristic polynomial to bedet(AtI).{\displaystyle \det(A-tI).} That polynomial differs from the one defined here by a sign(1)n,{\displaystyle (-1)^{n},} so it makes no difference for properties like having as roots the eigenvalues ofA{\displaystyle A}; however the definition above always gives amonic polynomial, whereas the alternative definition is monic only whenn{\displaystyle n} is even.

Examples

[edit]

To compute the characteristic polynomial of the matrixA=(2110).{\displaystyle A={\begin{pmatrix}2&1\\-1&0\end{pmatrix}}.}thedeterminant of the following is computed:tIA=(t211t0){\displaystyle tI-A={\begin{pmatrix}t-2&-1\\1&t-0\end{pmatrix}}} and found to be(t2)t1(1)=t22t+1,{\displaystyle (t-2)t-1(-1)=t^{2}-2t+1\,\!,} the characteristic polynomial ofA.{\displaystyle A.}

Another example useshyperbolic functions of ahyperbolic angle φ.For the matrix takeA=(cosh(φ)sinh(φ)sinh(φ)cosh(φ)).{\displaystyle A={\begin{pmatrix}\cosh(\varphi )&\sinh(\varphi )\\\sinh(\varphi )&\cosh(\varphi )\end{pmatrix}}.}Its characteristic polynomial isdet(tIA)=(tcosh(φ))2sinh2(φ)=t22t cosh(φ)+1=(teφ)(teφ).{\displaystyle \det(tI-A)=(t-\cosh(\varphi ))^{2}-\sinh ^{2}(\varphi )=t^{2}-2t\ \cosh(\varphi )+1=(t-e^{\varphi })(t-e^{-\varphi }).}

Properties

[edit]

The characteristic polynomialpA(t){\displaystyle p_{A}(t)} of an×n{\displaystyle n\times n} matrixA{\displaystyle A} ismonic (its leading coefficient is1{\displaystyle 1}) and its degree isn.{\displaystyle n.} The most important fact about the characteristic polynomial was already mentioned in themotivational paragraph: the eigenvalues ofA{\displaystyle A} are precisely theroots ofpA(t){\displaystyle p_{A}(t)} (this also holds for theminimal polynomial ofA,{\displaystyle A,} but its degree may be less thann{\displaystyle n}). All coefficients of the characteristic polynomial arepolynomial expressions in the entries of the matrix. In particular its constant coefficient oft0{\displaystyle t^{0}} isdet(A)=(1)ndet(A),{\displaystyle \det(-A)=(-1)^{n}\det(A),} the coefficient oftn{\displaystyle t^{n}} is 1, and the coefficient oftn1{\displaystyle t^{n-1}} istr(−A) = −tr(A), wheretr(A) is thetrace ofA.{\displaystyle A.} (The signs given here correspond to theformal definition given in the previous section; for the alternative definition these would instead bedet(A){\displaystyle \det(A)} and(−1)n – 1tr(A) respectively.[6])

For a2×2{\displaystyle 2\times 2} matrixA,{\displaystyle A,} the characteristic polynomial is thus given byt2tr(A)t+det(A).{\displaystyle t^{2}-\operatorname {tr} (A)t+\det(A).}

Using the language ofexterior algebra, the characteristic polynomial of ann×n{\displaystyle n\times n} matrixA{\displaystyle A} may be expressed aspA(t)=k=0ntnk(1)ktr(kA){\displaystyle p_{A}(t)=\sum _{k=0}^{n}t^{n-k}(-1)^{k}\operatorname {tr} \left(\textstyle \bigwedge ^{k}A\right)}wheretr(kA){\textstyle \operatorname {tr} \left(\bigwedge ^{k}A\right)} is thetrace of thek{\displaystyle k}thexterior power ofA,{\displaystyle A,} which has dimension(nk).{\textstyle {\binom {n}{k}}.} This trace may be computed as the sum of allprincipal minors ofA{\displaystyle A} of sizek.{\displaystyle k.} The recursiveFaddeev–LeVerrier algorithm computes these coefficients more efficiently[clarification needed].

When thecharacteristic of thefield of the coefficients is0,{\displaystyle 0,} each such trace may alternatively be computed as a single determinant, that of thek×k{\displaystyle k\times k} matrix,tr(kA)=1k!|trAk100trA2trAk20trAk1trAk21trAktrAk1trA| .{\displaystyle \operatorname {tr} \left(\textstyle \bigwedge ^{k}A\right)={\frac {1}{k!}}{\begin{vmatrix}\operatorname {tr} A&k-1&0&\cdots &0\\\operatorname {tr} A^{2}&\operatorname {tr} A&k-2&\cdots &0\\\vdots &\vdots &&\ddots &\vdots \\\operatorname {tr} A^{k-1}&\operatorname {tr} A^{k-2}&&\cdots &1\\\operatorname {tr} A^{k}&\operatorname {tr} A^{k-1}&&\cdots &\operatorname {tr} A\end{vmatrix}}~.}

TheCayley–Hamilton theorem states that replacingt{\displaystyle t} byA{\displaystyle A} in the characteristic polynomial (interpreting the resulting powers as matrix powers, and the constant termc{\displaystyle c} asc{\displaystyle c} times the identity matrix) yields thezero matrix. Informally speaking, every matrix satisfies its own characteristic equation. This statement is equivalent to saying that theminimal polynomial ofA{\displaystyle A} divides the characteristic polynomial ofA.{\displaystyle A.}

Twosimilar matrices have the same characteristic polynomial. The converse however is not true in general: two matrices with the same characteristic polynomial need not be similar.

The matrixA{\displaystyle A} and itstranspose have the same characteristic polynomial.A{\displaystyle A} is similar to atriangular matrixif and only if its characteristic polynomial can be completely factored into linear factors overK{\displaystyle K} (the same is true with the minimal polynomial instead of the characteristic polynomial). In this caseA{\displaystyle A} is similar to a matrix inJordan normal form.

Characteristic polynomial of a product of two matrices

[edit]

IfA{\displaystyle A} andB{\displaystyle B} are two squaren×n{\displaystyle n\times n} matrices then characteristic polynomials ofAB{\displaystyle AB} andBA{\displaystyle BA} coincide:pAB(t)=pBA(t).{\displaystyle p_{AB}(t)=p_{BA}(t).\,}

Proof: Ifλ{\displaystyle \lambda } is a non-zero generalized eigenvalue ofAB{\displaystyle AB} of algebraic multiplicityk{\displaystyle k}, andv{\displaystyle v} belongs to the kernel of(BAλ)k{\displaystyle (BA-\lambda )^{k}}, thenAv{\displaystyle Av} belongs to the kernel of(ABλ)k{\displaystyle (AB-\lambda )^{k}}, so the non-zero generalized eigenspaces ofAB{\displaystyle AB} andBA{\displaystyle BA} have the same dimension. Therefore, sinceAB{\displaystyle AB} andBA{\displaystyle BA} are bothn×n{\displaystyle n\times n}, the remaining generalized eigenspaces, with eigenvalue 0, have the same dimension. ThereforeAB{\displaystyle AB} andBA{\displaystyle BA} have the same characteristic polynomial, because all generalized eigenvalues are the same, with the same algebraic multiplicities.

More generally, ifA{\displaystyle A} is a matrix of orderm×n{\displaystyle m\times n} andB{\displaystyle B} is a matrix of ordern×m,{\displaystyle n\times m,} thenAB{\displaystyle AB} ism×m{\displaystyle m\times m} andBA{\displaystyle BA} isn×n{\displaystyle n\times n} matrix, and one haspBA(t)=tnmpAB(t).{\displaystyle p_{BA}(t)=t^{n-m}p_{AB}(t).\,}

To prove this, one may supposen>m,{\displaystyle n>m,} by exchanging, if needed,A{\displaystyle A} andB.{\displaystyle B.} Then, by borderingA{\displaystyle A} on the bottom bynm{\displaystyle n-m} rows of zeros, andB{\displaystyle B} on the right, by,nm{\displaystyle n-m} columns of zeros, one gets twon×n{\displaystyle n\times n} matricesA{\displaystyle A^{\prime }} andB{\displaystyle B^{\prime }} such thatBA=BA{\displaystyle B^{\prime }A^{\prime }=BA} andAB{\displaystyle A^{\prime }B^{\prime }} is equal toAB{\displaystyle AB} bordered bynm{\displaystyle n-m} rows and columns of zeros. The result follows from the case of square matrices, by comparing the characteristic polynomials ofAB{\displaystyle A^{\prime }B^{\prime }} andAB.{\displaystyle AB.}

Characteristic polynomial ofAk

[edit]

Ifλ{\displaystyle \lambda } is an eigenvalue of a square matrixA{\displaystyle A} with eigenvectorv,{\displaystyle \mathbf {v} ,} thenλk{\displaystyle \lambda ^{k}} is an eigenvalue ofAk{\displaystyle A^{k}} becauseAkv=Ak1Av=λAk1v==λkv.{\displaystyle A^{k}{\textbf {v}}=A^{k-1}A{\textbf {v}}=\lambda A^{k-1}{\textbf {v}}=\dots =\lambda ^{k}{\textbf {v}}.}

The multiplicities can be shown to agree as well, and this generalizes to any polynomial in place ofxk{\displaystyle x^{k}}:[7]

Theorem LetA{\displaystyle A} be a squaren×n{\displaystyle n\times n} matrix and letf(t){\displaystyle f(t)} be a polynomial. If the characteristic polynomial ofA{\displaystyle A} has a factorizationpA(t)=(tλ1)(tλ2)(tλn){\displaystyle p_{A}(t)=(t-\lambda _{1})(t-\lambda _{2})\cdots (t-\lambda _{n})}then the characteristic polynomial of the matrixf(A){\displaystyle f(A)} is given bypf(A)(t)=(tf(λ1))(tf(λ2))(tf(λn)).{\displaystyle p_{f(A)}(t)=(t-f(\lambda _{1}))(t-f(\lambda _{2}))\cdots (t-f(\lambda _{n})).}

That is, the algebraic multiplicity ofλ{\displaystyle \lambda } inf(A){\displaystyle f(A)} equals the sum of algebraic multiplicities ofλ{\displaystyle \lambda '} inA{\displaystyle A} overλ{\displaystyle \lambda '} such thatf(λ)=λ.{\displaystyle f(\lambda ')=\lambda .} In particular,tr(f(A))=i=1nf(λi){\displaystyle \operatorname {tr} (f(A))=\textstyle \sum _{i=1}^{n}f(\lambda _{i})} anddet(f(A))=i=1nf(λi).{\displaystyle \operatorname {det} (f(A))=\textstyle \prod _{i=1}^{n}f(\lambda _{i}).} Here a polynomialf(t)=t3+1,{\displaystyle f(t)=t^{3}+1,} for example, is evaluated on a matrixA{\displaystyle A} simply asf(A)=A3+I.{\displaystyle f(A)=A^{3}+I.}

The theorem applies to matrices and polynomials over any field orcommutative ring.[8]However, the assumption thatpA(t){\displaystyle p_{A}(t)} has a factorization into linear factors is not always true, unless the matrix is over analgebraically closed field such as the complex numbers.

Proof

This proof only applies to matrices and polynomials over complex numbers (or any algebraically closed field).In that case, the characteristic polynomial of any square matrix can be always factorized aspA(t)=(tλ1)(tλ2)(tλn){\displaystyle p_{A}(t)=\left(t-\lambda _{1}\right)\left(t-\lambda _{2}\right)\cdots \left(t-\lambda _{n}\right)}whereλ1,λ2,,λn{\displaystyle \lambda _{1},\lambda _{2},\ldots ,\lambda _{n}} are the eigenvalues ofA,{\displaystyle A,} possibly repeated. Moreover, theJordan decomposition theorem guarantees that any square matrixA{\displaystyle A} can be decomposed asA=S1US,{\displaystyle A=S^{-1}US,} whereS{\displaystyle S} is aninvertible matrix andU{\displaystyle U} isupper triangularwithλ1,,λn{\displaystyle \lambda _{1},\ldots ,\lambda _{n}} on the diagonal (with each eigenvalue repeated according to its algebraic multiplicity).(The Jordan normal form has stronger properties, but these are sufficient; alternatively theSchur decomposition can be used, which is less popular but somewhat easier to prove).

Letf(t)=iαiti.{\textstyle f(t)=\sum _{i}\alpha _{i}t^{i}.} Thenf(A)=αi(S1US)i=αiS1USS1USS1US=αiS1UiS=S1(αiUi)S=S1f(U)S.{\displaystyle f(A)=\textstyle \sum \alpha _{i}(S^{-1}US)^{i}=\textstyle \sum \alpha _{i}S^{-1}USS^{-1}US\cdots S^{-1}US=\textstyle \sum \alpha _{i}S^{-1}U^{i}S=S^{-1}(\textstyle \sum \alpha _{i}U^{i})S=S^{-1}f(U)S.} For an upper triangular matrixU{\displaystyle U} with diagonalλ1,,λn,{\displaystyle \lambda _{1},\dots ,\lambda _{n},} the matrixUi{\displaystyle U^{i}} is upper triangular with diagonalλ1i,,λni{\displaystyle \lambda _{1}^{i},\dots ,\lambda _{n}^{i}} inUi,{\displaystyle U^{i},} and hencef(U){\displaystyle f(U)} is upper triangular with diagonalf(λ1),,f(λn).{\displaystyle f\left(\lambda _{1}\right),\dots ,f\left(\lambda _{n}\right).} Therefore, the eigenvalues off(U){\displaystyle f(U)} aref(λ1),,f(λn).{\displaystyle f(\lambda _{1}),\dots ,f(\lambda _{n}).} Sincef(A)=S1f(U)S{\displaystyle f(A)=S^{-1}f(U)S} issimilar tof(U),{\displaystyle f(U),} it has the same eigenvalues, with the same algebraic multiplicities.

Secular function and secular equation

[edit]

Secular function

[edit]

The termsecular function has been used for what is now calledcharacteristic polynomial (in some literature the term secular function is still used). The term comes from the fact that the characteristic polynomial was used to calculatesecular perturbations (on a time scale of a century, that is, slow compared to annual motion) of planetary orbits, according toLagrange's theory of oscillations.

Secular equation

[edit]

Secular equation may have several meanings.

  • Inlinear algebra it is sometimes used in place of characteristic equation.
  • Inastronomy it is the algebraic or numerical expression of the magnitude of the inequalities in a planet's motion that remain after the inequalities of a short period have been allowed for.[9]

For general associative algebras

[edit]

The above definition of the characteristic polynomial of a matrixAMn(F){\displaystyle A\in M_{n}(F)} with entries in a fieldF{\displaystyle F} generalizes without any changes to the case whenF{\displaystyle F} is just acommutative ring.Garibaldi (2004) defines the characteristic polynomial for elements of an arbitrary finite-dimensional (associative, but not necessarily commutative) algebra over a fieldF{\displaystyle F} and proves the standard properties of the characteristic polynomial in this generality.

See also

[edit]

References

[edit]
  1. ^Guillemin, Ernst (1953).Introductory Circuit Theory. Wiley. pp. 366, 541.ISBN 0471330663.{{cite book}}:ISBN / Date incompatibility (help)
  2. ^Forsythe, George E.; Motzkin, Theodore (January 1952)."An Extension of Gauss' Transformation for Improving the Condition of Systems of Linear Equations"(PDF).Mathematics of Computation.6 (37):18–34.doi:10.1090/S0025-5718-1952-0048162-0. Retrieved3 October 2020.
  3. ^Frank, Evelyn (1946)."On the zeros of polynomials with complex coefficients"(PDF).Bulletin of the American Mathematical Society.52 (2):144–157.doi:10.1090/S0002-9904-1946-08526-2.
  4. ^"Characteristic Polynomial of a Graph – Wolfram MathWorld". RetrievedAugust 26, 2011.
  5. ^Steven Roman (1992).Advanced linear algebra (2 ed.). Springer. p. 137.ISBN 3540978372.
  6. ^Theorem 4 in theselecture notes
  7. ^Horn, Roger A.; Johnson, Charles R. (2013).Matrix Analysis (2nd ed.).Cambridge University Press. pp. 108–109, Section 2.4.2.ISBN 978-0-521-54823-6.
  8. ^Lang, Serge (1993).Algebra. New York: Springer. p.567, Theorem 3.10.ISBN 978-1-4613-0041-0.OCLC 852792828.
  9. ^"secular equation". RetrievedJanuary 21, 2010.
Retrieved from "https://en.wikipedia.org/w/index.php?title=Characteristic_polynomial&oldid=1321915547"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp