Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Trace (linear algebra)

From Wikipedia, the free encyclopedia
Sum of elements on the main diagonal
icon
This articleneeds additional citations forverification. Please helpimprove this article byadding citations to reliable sources. Unsourced material may be challenged and removed.
Find sources: "Trace" linear algebra – news ·newspapers ·books ·scholar ·JSTOR
(November 2023) (Learn how and when to remove this message)
This article needs editing tocomply with Wikipedia'sManual of Style. In particular, it has problems withMOS:FORMULA - avoid mixing<math>...</math> and{{math}} in the same expression. Please helpimprove the content.(July 2025) (Learn how and when to remove this message)

Inlinear algebra, thetrace of asquare matrixA, denotedtr(A),[1] is the sum of the elements on itsmain diagonal,a11+a22++ann{\displaystyle a_{11}+a_{22}+\dots +a_{nn}}. It is only defined for a square matrix (n ×n).

The trace of a matrix is the sum of itseigenvalues (counted with multiplicities). Also,tr(AB) = tr(BA) for any matricesA andB of the same size. Thus,similar matrices have the same trace. As a consequence, one can define the trace of alinear operator mapping a finite-dimensionalvector space into itself, since all matrices describing such an operator with respect to a basis are similar.

The trace is related to the derivative of thedeterminant (seeJacobi's formula).

Definition

[edit]

Thetrace of ann ×nsquare matrixA is defined as[1][2][3]: 34 tr(A)=i=1naii=a11+a22++ann{\displaystyle \operatorname {tr} (\mathbf {A} )=\sum _{i=1}^{n}a_{ii}=a_{11}+a_{22}+\dots +a_{nn}}whereaii denotes the entry on thei th row andi th column ofA. The entries ofA can bereal numbers,complex numbers, or more generally elements of afieldF. The trace is not defined for non-square matrices.

Example

[edit]

LetA be a matrix, withA=(a11a12a13a21a22a23a31a32a33)=(10311526125){\displaystyle \mathbf {A} ={\begin{pmatrix}a_{11}&a_{12}&a_{13}\\a_{21}&a_{22}&a_{23}\\a_{31}&a_{32}&a_{33}\end{pmatrix}}={\begin{pmatrix}1&0&3\\11&5&2\\6&12&-5\end{pmatrix}}}

Thentr(A)=i=13aii=a11+a22+a33=1+5+(5)=1{\displaystyle \operatorname {tr} (\mathbf {A} )=\sum _{i=1}^{3}a_{ii}=a_{11}+a_{22}+a_{33}=1+5+(-5)=1}

Properties

[edit]

Basic properties

[edit]

The trace is alinear mapping. That is,[1][2]tr(A+B)=tr(A)+tr(B)tr(cA)=ctr(A){\displaystyle {\begin{aligned}\operatorname {tr} (\mathbf {A} +\mathbf {B} )&=\operatorname {tr} (\mathbf {A} )+\operatorname {tr} (\mathbf {B} )\\\operatorname {tr} (c\mathbf {A} )&=c\operatorname {tr} (\mathbf {A} )\end{aligned}}}for all square matricesA andB, and allscalarsc.[3]: 34 

A matrix and itstranspose have the same trace:[1][2][3]: 34 tr(A)=tr(AT).{\displaystyle \operatorname {tr} (\mathbf {A} )=\operatorname {tr} \left(\mathbf {A} ^{\mathsf {T}}\right).}

This follows immediately from the fact that transposing a square matrix does not affect elements along the main diagonal.

Trace of a product

[edit]

The trace of a square matrix which is the product of two matrices can be rewritten as the sum of entry-wise products of their elements, i.e. as the sum of all elements of theirHadamard product. Phrased directly, ifA andB are twom ×n matrices, then:tr(ATB)=tr(ABT)=tr(BTA)=tr(BAT)=i=1mj=1naijbij.{\displaystyle \operatorname {tr} \left(\mathbf {A} ^{\mathsf {T}}\mathbf {B} \right)=\operatorname {tr} \left(\mathbf {A} \mathbf {B} ^{\mathsf {T}}\right)=\operatorname {tr} \left(\mathbf {B} ^{\mathsf {T}}\mathbf {A} \right)=\operatorname {tr} \left(\mathbf {B} \mathbf {A} ^{\mathsf {T}}\right)=\sum _{i=1}^{m}\sum _{j=1}^{n}a_{ij}b_{ij}\;.}

If one views any realm ×n matrix as a vector of lengthmn (an operation calledvectorization) then the above operation onA andB coincides with the standarddot product. According to the above expression,tr(AA) is a sum of squares and hence is nonnegative, equal to zero if and only ifA is zero.[4]: 7  Furthermore, as noted in the above formula,tr(AB) = tr(BA). These demonstrate the positive-definiteness and symmetry required of aninner product; it is common to calltr(AB) theFrobenius inner product ofA andB. This is a natural inner product on thevector space of all real matrices of fixed dimensions. Thenorm derived from this inner product is called theFrobenius norm, and it satisfies a submultiplicative property, as can be proven with theCauchy–Schwarz inequality:0[tr(AB)]2tr(ATA)tr(BTB),{\displaystyle 0\leq \left[\operatorname {tr} (\mathbf {A} \mathbf {B} )\right]^{2}\leq \operatorname {tr} \left(\mathbf {A} ^{\mathsf {T}}\mathbf {A} \right)\operatorname {tr} \left(\mathbf {B} ^{\mathsf {T}}\mathbf {B} \right),}ifA andB are real matrices such thatAB is a square matrix. The Frobenius inner product and norm arise frequently inmatrix calculus andstatistics.

The Frobenius inner product may be extended to ahermitian inner product on thecomplex vector space of all complex matrices of a fixed size, by replacingB by itscomplex conjugate.

The symmetry of the Frobenius inner product may be phrased more directly as follows: the matrices in the trace of a product can be switched without changing the result. IfA andB arem ×n andn ×m real or complex matrices, respectively, then[1][2][3]: 34 [note 1]

tr(AB)=tr(BA){\displaystyle \operatorname {tr} (\mathbf {A} \mathbf {B} )=\operatorname {tr} (\mathbf {B} \mathbf {A} )}

This is notable both for the fact thatAB does not usually equalBA, and also since the trace of either does not usually equaltr(A)tr(B).[note 2] Thesimilarity-invariance of the trace, meaning thattr(A) = tr(P−1AP) for any square matrixA and any invertible matrixP of the same dimensions, is a fundamental consequence. This is proved bytr(P1(AP))=tr((AP)P1)=tr(A).{\displaystyle \operatorname {tr} \left(\mathbf {P} ^{-1}(\mathbf {A} \mathbf {P} )\right)=\operatorname {tr} \left((\mathbf {A} \mathbf {P} )\mathbf {P} ^{-1}\right)=\operatorname {tr} (\mathbf {A} ).}Similarity invariance is the crucial property of the trace in order to discuss traces oflinear transformations as below.

Additionally, for real column vectorsaRn{\displaystyle \mathbf {a} \in \mathbb {R} ^{n}} andbRn{\displaystyle \mathbf {b} \in \mathbb {R} ^{n}}, the trace of the outer product is equivalent to the inner product:

tr(baT)=aTb{\displaystyle \operatorname {tr} \left(\mathbf {b} \mathbf {a} ^{\textsf {T}}\right)=\mathbf {a} ^{\textsf {T}}\mathbf {b} }

Cyclic property

[edit]

More generally, the trace isinvariant undercircular shifts, that is,

tr(ABCD)=tr(BCDA)=tr(CDAB)=tr(DABC).{\displaystyle \operatorname {tr} (\mathbf {A} \mathbf {B} \mathbf {C} \mathbf {D} )=\operatorname {tr} (\mathbf {B} \mathbf {C} \mathbf {D} \mathbf {A} )=\operatorname {tr} (\mathbf {C} \mathbf {D} \mathbf {A} \mathbf {B} )=\operatorname {tr} (\mathbf {D} \mathbf {A} \mathbf {B} \mathbf {C} ).}

This is known as thecyclic property.

Arbitrary permutations are not allowed: in general,tr(ABCD)tr(ACBD) .{\displaystyle \operatorname {tr} (\mathbf {A} \mathbf {B} \mathbf {C} \mathbf {D} )\neq \operatorname {tr} (\mathbf {A} \mathbf {C} \mathbf {B} \mathbf {D} )~.}

However, if products ofthreesymmetric matrices are considered, any permutation is allowed, since:tr(ABC)=tr((ABC)T)=tr(CBA)=tr(ACB),{\displaystyle \operatorname {tr} (\mathbf {A} \mathbf {B} \mathbf {C} )=\operatorname {tr} \left(\left(\mathbf {A} \mathbf {B} \mathbf {C} \right)^{\mathsf {T}}\right)=\operatorname {tr} (\mathbf {C} \mathbf {B} \mathbf {A} )=\operatorname {tr} (\mathbf {A} \mathbf {C} \mathbf {B} ),}where the first equality is because the traces of a matrix and its transpose are equal. Note that this is not true in general for more than three factors.

Trace of a Kronecker product

[edit]

The trace of theKronecker product of two matrices is the product of their traces:tr(AB)=tr(A)tr(B).{\displaystyle \operatorname {tr} (\mathbf {A} \otimes \mathbf {B} )=\operatorname {tr} (\mathbf {A} )\operatorname {tr} (\mathbf {B} ).}

Characterization of the trace

[edit]

The following three properties:tr(A+B)=tr(A)+tr(B),tr(cA)=ctr(A),tr(AB)=tr(BA),{\displaystyle {\begin{aligned}\operatorname {tr} (\mathbf {A} +\mathbf {B} )&=\operatorname {tr} (\mathbf {A} )+\operatorname {tr} (\mathbf {B} ),\\\operatorname {tr} (c\mathbf {A} )&=c\operatorname {tr} (\mathbf {A} ),\\\operatorname {tr} (\mathbf {A} \mathbf {B} )&=\operatorname {tr} (\mathbf {B} \mathbf {A} ),\end{aligned}}}characterize the traceup to a scalar multiple in the following sense: Iff{\displaystyle f} is alinear functional on the space of square matrices that satisfiesf(xy)=f(yx),{\displaystyle f(xy)=f(yx),} thenf{\displaystyle f} andtr{\displaystyle \operatorname {tr} } are proportional.[note 3]

Forn×n{\displaystyle n\times n} matrices, imposing the normalizationf(I)=n{\displaystyle f(\mathbf {I} )=n} makesf{\displaystyle f} equal to the trace.

Trace as the sum of eigenvalues

[edit]

Given anyn ×n matrixA, there is

tr(A)=i=1nλi{\displaystyle \operatorname {tr} (\mathbf {A} )=\sum _{i=1}^{n}\lambda _{i}}

whereλ1, ..., λn are theeigenvalues ofA counted with multiplicity. This holds true even ifA is a real matrix and some (or all) of the eigenvalues are complex numbers. This may be regarded as a consequence of the existence of theJordan canonical form, together with the similarity-invariance of the trace discussed above.

Trace of commutator

[edit]

When bothA andB aren ×n matrices, the trace of the (ring-theoretic)commutator ofA andB vanishes:tr([A,B]) = 0, becausetr(AB) = tr(BA) andtr is linear. One can state this as "the trace is a map ofLie algebrasglnk from operators to scalars", as the commutator of scalars is trivial (it is anAbelian Lie algebra). In particular, using similarity invariance, it follows that the identity matrix is never similar to the commutator of any pair of matrices.

Conversely, any square matrix with zero trace is a linear combination of the commutators of pairs of matrices.[note 4] Moreover, any square matrix with zero trace isunitarily equivalent to a square matrix with diagonal consisting of all zeros.

Traces of special kinds of matrices

[edit]

Relationship to the characteristic polynomial

[edit]

The trace of ann×n{\displaystyle n\times n} matrixA{\displaystyle A} is the coefficient oftn1{\displaystyle t^{n-1}} in thecharacteristic polynomial, possibly changed of sign, according to the convention in the definition of the characteristic polynomial.

Relationship to eigenvalues

[edit]

IfA is a linear operator represented by a square matrix withreal orcomplex entries and ifλ1, ...,λn are theeigenvalues ofA (listed according to theiralgebraic multiplicities), then

tr(A)=iλi{\displaystyle \operatorname {tr} (\mathbf {A} )=\sum _{i}\lambda _{i}}

This follows from the fact thatA is alwayssimilar to itsJordan form, an uppertriangular matrix havingλ1, ...,λn on the main diagonal. In contrast, thedeterminant ofA is theproduct of its eigenvalues; that is,det(A)=iλi.{\displaystyle \det(\mathbf {A} )=\prod _{i}\lambda _{i}.}

Everything in the present section applies as well to any square matrix with coefficients in analgebraically closed field.

Derivative relationships

[edit]

Ifa is a square matrixwith small entries andI denotes theidentity matrix, then we have approximately

det(I+a)1+tr(a).{\displaystyle \det(\mathbf {I} +\mathbf {a} )\approx 1+\operatorname {tr} (\mathbf {a} ).}

Precisely this means that the trace is thederivative of thedeterminant function at the identity matrix.Jacobi's formula

ddet(A)=tr(adj(A)dA){\displaystyle d\det(\mathbf {A} )=\operatorname {tr} {\big (}\operatorname {adj} (\mathbf {A} )\cdot d\mathbf {A} {\big )}}

is more general and describes thedifferential of the determinant at an arbitrary square matrix, in terms of the trace and theadjugate of the matrix.

From this (or from the connection between the trace and the eigenvalues), one can derive a relation between the trace function, thematrix exponential function, and the determinant:det(exp(A))=exp(tr(A)).{\displaystyle \det(\exp(\mathbf {A} ))=\exp(\operatorname {tr} (\mathbf {A} )).}

A related characterization of the trace applies to linearvector fields. Given a matrixA, define a vector fieldF onRn byF(x) =Ax. The components of this vector field are linear functions (given by the rows ofA). ItsdivergencedivF is a constant function, whose value is equal totr(A).

By thedivergence theorem, one can interpret this in terms of flows: ifF(x) represents the velocity of a fluid at locationx andU is a region inRn, thenet flow of the fluid out ofU is given bytr(A) · vol(U), wherevol(U) is thevolume ofU.

The trace is a linear operator, hence it commutes with the derivative:dtr(X)=tr(dX).{\displaystyle d\operatorname {tr} (\mathbf {X} )=\operatorname {tr} (d\mathbf {X} ).}

Trace of a linear operator

[edit]

In general, given some linear mapf :VV of finiterank (whereV is avector space), we can define the trace of this map by considering the trace of amatrix representation off, that is, choosing abasis forV and describingf as a matrix relative to this basis, and taking the trace of this square matrix. The result will not depend on the basis chosen, since different bases will give rise tosimilar matrices, allowing for the possibility of a basis-independent definition for the trace of a linear map.

Such a definition can be given using thecanonical isomorphism between the space of linear endomorphisms ofV of finiterank andVV*, whereV* is thedual space ofV. Letv be inV and letg be inV*. Then the trace of the decomposable elementvg is defined to beg(v); the trace of a general element is defined by linearity. The trace of a linear mapf :VV of finite rank can then be defined as the trace, in the above sense, of the element ofVV* corresponding tof under the above-mentioned canonical isomorphism. Using an explicit basis forV and the corresponding dual basis forV*, one can show that this gives the same definition of the trace as given above.

Numerical algorithms

[edit]

Stochastic estimator

[edit]

The trace can be estimated unbiasedly by "Hutchinson's trick":[5]

Given any matrixWRn×n{\displaystyle {\boldsymbol {W}}\in \mathbb {R} ^{n\times n}}, and any randomuRn{\displaystyle {\boldsymbol {u}}\in \mathbb {R} ^{n}} withE[uu]=I{\displaystyle \mathbb {E} [{\boldsymbol {u}}{\boldsymbol {u}}^{\intercal }]=\mathbf {I} }, we haveE[uWu]=trW{\displaystyle \mathbb {E} [{\boldsymbol {u}}^{\intercal }{\boldsymbol {W}}{\boldsymbol {u}}]=\operatorname {tr} {\boldsymbol {W}}}.

For a proof expand the expectation directly.

Usually, the random vector is sampled fromN(0,I){\displaystyle \operatorname {N} (\mathbf {0} ,\mathbf {I} )} (normal distribution) or{±n1/2}n{\displaystyle \{\pm n^{-1/2}\}^{n}} (Rademacher distribution).

More sophisticated stochastic estimators of trace have been developed.[6]

Applications

[edit]

If a 2 x 2 real matrix has zero trace, its square is adiagonal matrix.

The trace of a 2 × 2complex matrix is used to classifyMöbius transformations. First, the matrix is normalized to make itsdeterminant equal to one. Then, if the square of the trace is 4, the corresponding transformation isparabolic. If the square is in the interval[0,4), it iselliptic. Finally, if the square is greater than 4, the transformation isloxodromic. Seeclassification of Möbius transformations.

The trace is used to definecharacters ofgroup representations. Two representationsA,B :GGL(V) of a groupG are equivalent (up to change of basis onV) iftr(A(g)) = tr(B(g)) for allgG.

The trace also plays a central role in the distribution ofquadratic forms.

Lie algebra

[edit]

The trace is a map of Lie algebrastr:glnK{\displaystyle \operatorname {tr} :{\mathfrak {gl}}_{n}\to K} from the Lie algebragln{\displaystyle {\mathfrak {gl}}_{n}} of linear operators on ann-dimensional space (n ×n matrices with entries inK{\displaystyle K}) to the Lie algebraK of scalars; asK is Abelian (the Lie bracket vanishes), the fact that this is a map of Lie algebras is exactly the statement that the trace of a bracket vanishes:tr([A,B])=0 for each A,Bgln.{\displaystyle \operatorname {tr} ([\mathbf {A} ,\mathbf {B} ])=0{\text{ for each }}\mathbf {A} ,\mathbf {B} \in {\mathfrak {gl}}_{n}.}

The kernel of this map, a matrix whose trace iszero, is often said to betraceless ortrace free, and these matrices form thesimple Lie algebrasln{\displaystyle {\mathfrak {sl}}_{n}}, which is theLie algebra of thespecial linear group of matrices with determinant 1. The special linear group consists of the matrices which do not change volume, while thespecial linear Lie algebra is the matrices which do not alter volume ofinfinitesimal sets.

In fact, there is an internaldirect sum decompositiongln=slnK{\displaystyle {\mathfrak {gl}}_{n}={\mathfrak {sl}}_{n}\oplus K} of operators/matrices into traceless operators/matrices and scalars operators/matrices. The projection map onto scalar operators can be expressed in terms of the trace, concretely as:A1ntr(A)I.{\displaystyle \mathbf {A} \mapsto {\frac {1}{n}}\operatorname {tr} (\mathbf {A} )\mathbf {I} .}

Formally, one can compose the trace (thecounit map) with the unit mapKgln{\displaystyle K\to {\mathfrak {gl}}_{n}} of "inclusion ofscalars" to obtain a mapglngln{\displaystyle {\mathfrak {gl}}_{n}\to {\mathfrak {gl}}_{n}} mapping onto scalars, and multiplying byn. Dividing byn makes this a projection, yielding the formula above.

In terms ofshort exact sequences, one has0slnglntrK0{\displaystyle 0\to {\mathfrak {sl}}_{n}\to {\mathfrak {gl}}_{n}{\overset {\operatorname {tr} }{\to }}K\to 0}which is analogous to1SLnGLndetK1{\displaystyle 1\to \operatorname {SL} _{n}\to \operatorname {GL} _{n}{\overset {\det }{\to }}K^{*}\to 1}(whereK=K{0}{\displaystyle K^{*}=K\setminus \{0\}}) forLie groups. However, the trace splits naturally (via1/n{\displaystyle 1/n} times scalars) sogln=slnK{\displaystyle {\mathfrak {gl}}_{n}={\mathfrak {sl}}_{n}\oplus K}, but the splitting of the determinant would be as thenth root times scalars, and this does not in general define a function, so the determinant does not split and thegeneral linear group does not decompose:GLnSLn×K.{\displaystyle \operatorname {GL} _{n}\neq \operatorname {SL} _{n}\times K^{*}.}

Bilinear forms

[edit]

Thebilinear form (whereX,Y are square matrices)B(X,Y)=tr(ad(X)ad(Y)){\displaystyle B(\mathbf {X} ,\mathbf {Y} )=\operatorname {tr} (\operatorname {ad} (\mathbf {X} )\operatorname {ad} (\mathbf {Y} ))}

wheread(X)Y=[X,Y]=XYYX{\displaystyle \operatorname {ad} (\mathbf {X} )\mathbf {Y} =[\mathbf {X} ,\mathbf {Y} ]=\mathbf {X} \mathbf {Y} -\mathbf {Y} \mathbf {X} }
and for orientation, ifdetY0{\displaystyle \operatorname {det} \mathbf {Y} \neq 0}
thenad(X)=XYXY1 .{\displaystyle \operatorname {ad} (\mathbf {X} )=\mathbf {X} -\mathbf {Y} \mathbf {X} \mathbf {Y} ^{-1}~.}

B(X,Y){\displaystyle B(\mathbf {X} ,\mathbf {Y} )} is called theKilling form; it is used to classifyLie algebras.

The trace defines a bilinear form:(X,Y)tr(XY) .{\displaystyle (\mathbf {X} ,\mathbf {Y} )\mapsto \operatorname {tr} (\mathbf {X} \mathbf {Y} )~.}

The form is symmetric, non-degenerate[note 5] and associative in the sense that:tr(X[Y,Z])=tr([X,Y]Z).{\displaystyle \operatorname {tr} (\mathbf {X} [\mathbf {Y} ,\mathbf {Z} ])=\operatorname {tr} ([\mathbf {X} ,\mathbf {Y} ]\mathbf {Z} ).}

For a complex simple Lie algebra (such assl{\displaystyle {\mathfrak {sl}}}n), every such bilinear form is proportional to each other; in particular, to the Killing form[citation needed].

Two matricesX andY are said to betrace orthogonal iftr(XY)=0.{\displaystyle \operatorname {tr} (\mathbf {X} \mathbf {Y} )=0.}

There is a generalization to a general representation(ρ,g,V){\displaystyle (\rho ,{\mathfrak {g}},V)} of a Lie algebrag{\displaystyle {\mathfrak {g}}}, such thatρ{\displaystyle \rho } is a homomorphism of Lie algebrasρ:gEnd(V).{\displaystyle \rho :{\mathfrak {g}}\rightarrow {\text{End}}(V).} The trace formtrV{\displaystyle {\text{tr}}_{V}} onEnd(V){\displaystyle {\text{End}}(V)} is defined as above. The bilinear formϕ(X,Y)=trV(ρ(X)ρ(Y)){\displaystyle \phi (\mathbf {X} ,\mathbf {Y} )={\text{tr}}_{V}(\rho (\mathbf {X} )\rho (\mathbf {Y} ))}is symmetric and invariant due to cyclicity.

Generalizations

[edit]

The concept of trace of a matrix is generalized to thetrace class ofcompact operators onHilbert spaces, and the analog of theFrobenius norm is called theHilbert–Schmidt norm.

IfK{\displaystyle K} is a trace-class operator, then for anyorthonormal basis{en}n=1{\displaystyle \{e_{n}\}_{n=1}}, the trace is given bytr(K)=nen,Ken,{\displaystyle \operatorname {tr} (K)=\sum _{n}\left\langle e_{n},Ke_{n}\right\rangle ,}and is finite and independent of the orthonormal basis.[7]

Thepartial trace is another generalization of the trace that is operator-valued. The trace of a linear operatorZ{\displaystyle Z} which lives on a product spaceAB{\displaystyle A\otimes B} is equal to the partial traces overA{\displaystyle A} andB{\displaystyle B}:tr(Z)=trA(trB(Z))=trB(trA(Z)).{\displaystyle \operatorname {tr} (Z)=\operatorname {tr} _{A}\left(\operatorname {tr} _{B}(Z)\right)=\operatorname {tr} _{B}\left(\operatorname {tr} _{A}(Z)\right).}

For more properties and a generalization of the partial trace, seetraced monoidal categories.

IfA{\displaystyle A} is a generalassociative algebra over a fieldk{\displaystyle k}, then a trace onA{\displaystyle A} is often defined to be anyfunctionaltr:Ak{\displaystyle \operatorname {tr} :A\to k} which vanishes on commutators;tr([a,b])=0{\displaystyle \operatorname {tr} ([a,b])=0} for alla,bA{\displaystyle a,b\in A}. Such a trace is not uniquely defined; it can always at least be modified by multiplication by a nonzero scalar.

Asupertrace is the generalization of a trace to the setting ofsuperalgebras.

The operation oftensor contraction generalizes the trace to arbitrary tensors.

Gomme and Klein (2011) define a matrix trace operatortrm{\displaystyle \operatorname {trm} } that operates onblock matrices and use it to compute second-order perturbation solutions to dynamic economic models without the need fortensor notation.[8]

Traces in the language of tensor products

[edit]

Given a vector spaceV, there is a natural bilinear mapV ×VF given by sending(v, φ) to the scalarφ(v). Theuniversal property of thetensor productVV automatically implies that this bilinear map is induced by a linear functional onVV.[9]

Similarly, there is a natural bilinear mapV ×V → Hom(V,V) given by sending(v, φ) to the linear mapw ↦ φ(w)v. The universal property of the tensor product, just as used previously, says that this bilinear map is induced by a linear mapVV → Hom(V,V). IfV is finite-dimensional, then this linear map is alinear isomorphism.[9] This fundamental fact is a straightforward consequence of the existence of a (finite) basis ofV, and can also be phrased as saying that any linear mapVV can be written as the sum of (finitely many) rank-one linear maps. Composing the inverse of the isomorphism with the linear functional obtained above results in a linear functional onHom(V,V). This linear functional is exactly the same as the trace.

Using the definition of trace as the sum of diagonal elements, the matrix formulatr(AB) = tr(BA) is straightforward to prove, and was given above. In the present perspective, one is considering linear mapsS andT, and viewing them as sums of rank-one maps, so that there are linear functionalsφi andψj and nonzero vectorsvi andwj such thatS(u) = Σφi(u)vi andT(u) = Σψj(u)wj for anyu inV. Then

(ST)(u)=iφi(jψj(u)wj)vi=ijψj(u)φi(wj)vi{\displaystyle (S\circ T)(u)=\sum _{i}\varphi _{i}\left(\sum _{j}\psi _{j}(u)w_{j}\right)v_{i}=\sum _{i}\sum _{j}\psi _{j}(u)\varphi _{i}(w_{j})v_{i}}

for anyu inV. The rank-one linear mapuψj(u)φi(wj)vi has traceψj(vi)φi(wj) and so

tr(ST)=ijψj(vi)φi(wj)=jiφi(wj)ψj(vi).{\displaystyle \operatorname {tr} (S\circ T)=\sum _{i}\sum _{j}\psi _{j}(v_{i})\varphi _{i}(w_{j})=\sum _{j}\sum _{i}\varphi _{i}(w_{j})\psi _{j}(v_{i}).}

Following the same procedure withS andT reversed, one finds exactly the same formula, proving thattr(ST) equalstr(TS).

The above proof can be regarded as being based upon tensor products, given that the fundamental identity ofEnd(V) withVV is equivalent to the expressibility of any linear map as the sum of rank-one linear maps. As such, the proof may be written in the notation of tensor products. Then one may consider the multilinear mapV ×V ×V ×VVV given by sending(v,φ,w,ψ) toφ(w)vψ. Further composition with the trace map then results inφ(w)ψ(v), and this is unchanged if one were to have started with(w,ψ,v,φ) instead. One may also consider the bilinear mapEnd(V) × End(V) → End(V) given by sending(f,g) to the compositionfg, which is then induced by a linear mapEnd(V) ⊗ End(V) → End(V). It can be seen that this coincides with the linear mapVVVVVV. The established symmetry upon composition with the trace map then establishes the equality of the two traces.[9]

For any finite dimensional vector spaceV, there is a natural linear mapFVV'; in the language of linear maps, it assigns to a scalarc the linear mapc⋅idV. Sometimes this is calledcoevaluation map, and the traceVV'F is calledevaluation map.[9] These structures can be axiomatized to definecategorical traces in the abstract setting ofcategory theory.

See also

[edit]

Notes

[edit]
  1. ^This is immediate from the definition of thematrix product:tr(AB)=i=1m(AB)ii=i=1mj=1naijbji=j=1ni=1mbjiaij=j=1n(BA)jj=tr(BA).{\displaystyle \operatorname {tr} (\mathbf {A} \mathbf {B} )=\sum _{i=1}^{m}\left(\mathbf {A} \mathbf {B} \right)_{ii}=\sum _{i=1}^{m}\sum _{j=1}^{n}a_{ij}b_{ji}=\sum _{j=1}^{n}\sum _{i=1}^{m}b_{ji}a_{ij}=\sum _{j=1}^{n}\left(\mathbf {B} \mathbf {A} \right)_{jj}=\operatorname {tr} (\mathbf {B} \mathbf {A} ).}
  2. ^For example, ifA=(0100),B=(0010),{\displaystyle \mathbf {A} ={\begin{pmatrix}0&1\\0&0\end{pmatrix}},\quad \mathbf {B} ={\begin{pmatrix}0&0\\1&0\end{pmatrix}},}then the product isAB=(1000),{\displaystyle \mathbf {AB} ={\begin{pmatrix}1&0\\0&0\end{pmatrix}},}and the traces aretr(AB) = 1 ≠ 0 ⋅ 0 = tr(A)tr(B).
  3. ^Proof: Leteij{\displaystyle e_{ij}} the standard basis and note thatf(eij)=f(eiej)=f(eie1e1ej)=f(e1ejeie1)=f(0)=0{\displaystyle f\left(e_{ij}\right)=f\left(e_{i}e_{j}^{\top }\right)=f\left(e_{i}e_{1}^{\top }e_{1}e_{j}^{\top }\right)=f\left(e_{1}e_{j}^{\top }e_{i}e_{1}^{\top }\right)=f\left(0\right)=0} ifij{\displaystyle i\neq j} andf(ejj)=f(e11){\displaystyle f\left(e_{jj}\right)=f\left(e_{11}\right)}f(A)=i,j[A]ijf(eij)=i[A]iif(e11)=f(e11)tr(A).{\displaystyle f(\mathbf {A} )=\sum _{i,j}[\mathbf {A} ]_{ij}f\left(e_{ij}\right)=\sum _{i}[\mathbf {A} ]_{ii}f\left(e_{11}\right)=f\left(e_{11}\right)\operatorname {tr} (\mathbf {A} ).}More abstractly, this corresponds to the decompositiongln=slnk,{\displaystyle {\mathfrak {gl}}_{n}={\mathfrak {sl}}_{n}\oplus k,}astr(AB)=tr(BA){\displaystyle \operatorname {tr} (AB)=\operatorname {tr} (BA)} (equivalently,tr([A,B])=0{\displaystyle \operatorname {tr} ([A,B])=0}) defines the trace onsln,{\displaystyle {\mathfrak {sl}}_{n},} which has complement the scalar matrices, and leaves one degree of freedom: any such map is determined by its value on scalars, which is one scalar parameter and hence all are multiple of the trace, a nonzero such map.
  4. ^Proof:sln{\displaystyle {\mathfrak {sl}}_{n}} is asemisimple Lie algebra and thus every element in it is a linear combination of commutators of some pairs of elements, otherwise thederived algebra would be a proper ideal.
  5. ^This follows from the fact thattr(A*A) = 0if and only ifA =0.

References

[edit]
  1. ^abcde"Rank, trace, determinant, transpose, and inverse of matrices".fourier.eng.hmc.edu. Archived fromthe original on 2019-07-01. Retrieved2020-09-09.
  2. ^abcdWeisstein, Eric W. (2003) [1999]."Trace (matrix)". In Weisstein, Eric W. (ed.).CRC Concise Encyclopedia of Mathematics (2nd ed.). Boca Raton, FL:Chapman & Hall.doi:10.1201/9781420035223.ISBN 1-58488-347-2.MR 1944431.Zbl 1079.00009. Retrieved2020-09-09.
  3. ^abcdLipschutz, Seymour; Lipson, Marc (September 2005).Theory and Problems of Linear Algebra. Schaum's Outline. McGraw-Hill.ISBN 9780070605022.
  4. ^Horn, Roger A.; Johnson, Charles R. (2013).Matrix Analysis (2nd ed.). Cambridge University Press.ISBN 9780521839402.
  5. ^Hutchinson, M.F. (January 1989)."A Stochastic Estimator of the Trace of the Influence Matrix for Laplacian Smoothing Splines".Communications in Statistics - Simulation and Computation.18 (3):1059–1076.doi:10.1080/03610918908812806.ISSN 0361-0918.
  6. ^Avron, Haim; Toledo, Sivan (2011-04-11)."Randomized algorithms for estimating the trace of an implicit symmetric positive semi-definite matrix".Journal of the ACM.58 (2): 8:1–8:34.doi:10.1145/1944345.1944349.ISSN 0004-5411.S2CID 5827717.
  7. ^Teschl, G. (30 October 2014).Mathematical Methods in Quantum Mechanics. Graduate Studies in Mathematics. Vol. 157 (2nd ed.). American Mathematical Society.ISBN 978-1470417048.
  8. ^P. Gomme, P. Klein (2011). "Second-order approximation of dynamic models without the use of tensors".Journal of Economic Dynamics & Control.35 (4):604–615.doi:10.1016/j.jedc.2010.10.006.
  9. ^abcdKassel, Christian (1995).Quantum groups.Graduate Texts in Mathematics. Vol. 155. New York:Springer-Verlag.doi:10.1007/978-1-4612-0783-2.ISBN 0-387-94370-6.MR 1321145.Zbl 0808.17003.

External links

[edit]
Linear equations
Three dimensional Euclidean space
Matrices
Matrix decompositions
Relations and computations
Vector spaces
Structures
Multilinear algebra
Affine and projective
Numerical linear algebra
Retrieved from "https://en.wikipedia.org/w/index.php?title=Trace_(linear_algebra)&oldid=1335191830"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp