Inmathematics, thesquare root of a matrix extends the notion ofsquare root from numbers tomatrices. A matrixB is said to be a square root ofA if thematrix productBB is equal toA.[1]
Some authors use the namesquare root or the notationA1/2 only for the specific case whenA ispositive semidefinite, to denote the unique matrixB that is positive semidefinite and such thatBB =BTB =A (for real-valued matrices, whereBT is thetranspose ofB).
Less frequently, the namesquare root may be used for any factorization of a positive semidefinite matrixA asBTB =A, as in theCholesky factorization, even ifBB ≠ A. This distinct meaning is discussed inPositive definite matrix § Decomposition.
In general, a matrix can have several square roots. In particular, if then as well.
For example, the 2×2identity matrix has infinitely many square roots. They are given by
where are any numbers (real or complex) such thatIn particular if is anyPythagorean triple[a] then is one of the matrix square roots of which happens to be symmetric and has rational entries.[2]Thus
MinusI2 also has a square root, for example:
which can be used to represent theimaginary unit and hence allcomplex numbers using 2×2 real matrices, seematrix representation of complex numbers.
Just as with thereal numbers, a real matrix may fail to have a real square root, but have a square root withcomplex-valued entries.Some matrices have no square root. An example is the matrix
Notice that some ideas fromnumber theory do not carry over to matrices: The square root of a nonnegativeinteger must either be another integer or anirrational number, excluding non-integerrationals. Contrast that to a matrix of integers, which can have a square root whose entries are all non-integer rational numbers, as demonstrated in some of the above examples.
A symmetric realn ×n matrix is calledpositive semidefinite if for all (here denotes thetranspose, changing a column vectorx into a row vector).A square real matrix is positive semidefinite if and only if for some matrixB.There can be many different such matricesB.A positive semidefinite matrixA can also have many matricesB such that.However,A always has precisely one square rootB that isboth positive semidefinite and symmetric.In particular, sinceB is required to be symmetric,, so the two conditions or are equivalent.
For complex-valued matrices, theconjugate transpose is used instead and positive semidefinite matrices areHermitian, meaning.
Theorem[3]— LetA be a positive semidefinite matrix that is also symmetric.[b] Then there is exactly one positive semidefinite and symmetric matrixB such that.[c]
This unique matrix is called theprincipal,non-negative, orpositive square root (the latter in the case ofpositive definite matrices).
The principal square root of a real positive semidefinite matrix is real.[3]The principal square root of a positive definite matrix is positive definite; more generally, the rank of the principal square root ofA is the same as the rank ofA.[3]
The operation of taking the principal square root is continuous on this set of matrices.[4] These properties are consequences of theholomorphic functional calculus applied to matrices.[5][6]The existence and uniqueness of the principal square root can be deduced directly from theJordan normal form (see below).
Ann×n matrix withndistinct nonzero eigenvalues has 2n square roots. Such a matrix,A, has aneigendecompositionVDV−1 whereV is the matrix whose columns are eigenvectors ofA andD is the diagonal matrix whose diagonal elements are the correspondingn eigenvaluesλi. Thus the square roots ofA are given byVD1/2V−1, whereD1/2 is any square root matrix ofD, which, for distinct eigenvalues, must be diagonal with diagonal elements equal to square roots of the diagonal elements ofD; since there are two possible choices for a square root of each diagonal element ofD, there are 2n choices for the matrixD1/2.
This also leads to a proof of the above observation, that a positive-definite matrix has precisely one positive-definite square root: a positive definite matrix has only positive eigenvalues, and each of these eigenvalues has only one positive square root; and since the eigenvalues of the square root matrix are the diagonal elements ofD1/2, for the square root matrix to be itself positive definite necessitates the use of only the unique positive square roots of the original eigenvalues.
If a matrix isidempotent, meaning, then by definition one of its square roots is the matrix itself.
IfD is adiagonaln ×n matrix,then some of its square roots are diagonal matrices, where.If the diagonal elements ofD are real and non-negative then it is positive semidefinite, and if the square roots are taken with the (+) sign (i.e. all non-negative), the resulting matrix is the principal root ofD.A diagonal matrix may have additional non-diagonal roots if some entries on the diagonal are equal, as exemplified by the identity matrix above.
IfU is anupper triangular matrix (meaning its entries are for) and at most one of its diagonal entries is zero, then one upper triangular solution of the equation can be found as follows.Since the equation should be satisfied, let be theprincipal square root of the complex number.By the assumption, this guarantees that for alli,j (because the principal square roots of complex numbers all lie on one half of the complex plane).From the equation
we deduce that can be computed recursively for increasing from 1 ton-1 as:
IfU is upper triangular but has multiple zeroes on the diagonal, then a square root might not exist, as exemplified by.Note the diagonal entries of a triangular matrix are precisely itseigenvalues (seeTriangular matrix#Properties).
Ann ×n matrixA isdiagonalizable if there is a matrixV and a diagonal matrixD such thatA =VDV−1. This happens if and only ifA hasneigenvectors which constitute a basis forCn. In this case,V can be chosen to be the matrix with then eigenvectors as columns, and thus a square root ofA is
whereS is any square root ofD. Indeed,
For example, the matrix can be diagonalized asVDV−1, where
D has principal square root
giving the square root
WhenA is symmetric, the diagonalizing matrixV can be made anorthogonal matrix by suitably choosing the eigenvectors (seespectral theorem). Then the inverse ofV is simply the transpose, so that
Every complex-valued square matrix, regardless of diagonalizability, has aSchur decomposition given by where is upper triangular and isunitary (meaning).Theeigenvalues ofare exactly the diagonal entries of;if at most one of them is zero, then the following is a square root[7]
where a square root of the upper triangular matrix can be found as described above.
If is positive definite, then the eigenvalues are all positive reals, so the chosen diagonal of also consists of positive reals.Hence the eigenvalues of are positive reals, which means the resulting matrix is the principal root of.
Similarly as for the Schur decomposition, every square matrix can be decomposed as whereP isinvertible andJ is inJordan normal form.
To see that any complex matrix with positive eigenvalues has a square root of the same form, it suffices to check this for a Jordan block. Any such block has the form λ(I +N) with λ > 0 andNnilpotent. If(1 +z)1/2 = 1 +a1z +a2z2 + ⋯ is the binomial expansion for the square root (valid in |z| < 1), then as aformal power series its square equals 1 +z. SubstitutingN forz, only finitely many terms will be non-zero andS = √λ (I +a1N +a2N2 + ⋯) gives a square root of the Jordan block with eigenvalue√λ.
It suffices to check uniqueness for a Jordan block with λ = 1. The square constructed above has the formS =I +L whereL is polynomial inN without constant term. Any other square rootT with positive eigenvalues has the formT =I +M withM nilpotent, commuting withN and henceL. But then0 =S2 −T2 = 2(L −M)(I + (L +M)/2). SinceL andM commute, the matrixL +M is nilpotent andI + (L +M)/2 is invertible with inverse given by aNeumann series. HenceL =M.
IfA is a matrix with positive eigenvalues andminimal polynomialp(t), then the Jordan decomposition into generalized eigenspaces ofA can be deduced from the partial fraction expansion ofp(t)−1. The corresponding projections onto the generalized eigenspaces are given by real polynomials inA. On each eigenspace,A has the formλ(I +N) as above. Thepower series expression for the square root on the eigenspace show that the principal square root ofA has the formq(A) whereq(t) is a polynomial with real coefficients.
Recall the formal power series, which converges provided (since the coefficients of the power series are summable). Plugging in into this expression yields
provided that. By virtue ofGelfand formula, that condition is equivalent to the requirement that the spectrum of is contained within the disk. This method of defining or computing is especially useful in the case where is positive semi-definite. In that case, we have and therefore, so that the expression defines a square root of which moreover turns out to be the unique positive semi-definite root. This method remains valid to define square roots of operators on infinite-dimensional Banach or Hilbert spaces or certain elements of (C*) Banach algebras.
Another way to find the square root of ann ×n matrixA is the Denman–Beavers square root iteration.[8]
LetY0 =A andZ0 =I, whereI is then ×nidentity matrix. The iteration is defined by
As this uses a pair of sequences of matrix inverses whose later elements change comparatively little, only the first elements have a high computational cost since the remainder can be computed from earlier elements with only a few passes of a variant ofNewton's method forcomputing inverses,
With this, for later values ofk one would set and and then use for some small (perhaps just 1), and similarly for
Convergence is not guaranteed, even for matrices that do have square roots, but if the process converges, the matrix converges quadratically to a square rootA1/2, while converges to its inverse,A−1/2.
Yet another iterative method is obtained by taking the well-known formula of theBabylonian method for computing the square root of a real number, and applying it to matrices. LetX0 =I, whereI is theidentity matrix. The iteration is defined by
Again, convergence is not guaranteed, but if the process converges, the matrix converges quadratically to a square rootA1/2. Compared to Denman–Beavers iteration, an advantage of the Babylonian method is that only onematrix inverse need be computed per iteration step. On the other hand, as Denman–Beavers iteration uses a pair of sequences of matrix inverses whose later elements change comparatively little, only the first elements have a high computational cost since the remainder can be computed from earlier elements with only a few passes of a variant ofNewton's method forcomputing inverses (seeDenman–Beavers iteration above); of course, the same approach can be used to get the single sequence of inverses needed for the Babylonian method. However, unlike Denman–Beavers iteration, the Babylonian method is numerically unstable and more likely to fail to converge.[1]
The Babylonian method follows fromNewton's method for the equation and using forall[9]
This articleneeds additional citations forverification. Please helpimprove this article byadding citations to reliable sources. Unsourced material may be challenged and removed. Find sources: "Square root of a matrix" – news ·newspapers ·books ·scholar ·JSTOR(July 2010) (Learn how and when to remove this message) |
Inlinear algebra andoperator theory, given aboundedpositive semidefinite operator (a non-negative operator)T on a complex Hilbert space,B is a square root ofT ifT =B* B, whereB* denotes theHermitian adjoint ofB.[citation needed] According to thespectral theorem, thecontinuous functional calculus can be applied to obtain an operatorT1/2 such thatT1/2 is itself positive and (T1/2)2 =T. The operatorT1/2 is theunique non-negative square root ofT.[citation needed]
A bounded non-negative operator on a complex Hilbert space is self adjoint by definition. SoT = (T1/2)*T1/2. Conversely, it is trivially true that every operator of the formB* B is non-negative. Therefore, an operatorT is non-negativeif and only ifT =B* B for someB (equivalently,T =CC* for someC).
TheCholesky factorization provides another particular example of square root, which should not be confused with the unique non-negative square root.
IfT is a non-negative operator on a finite-dimensional Hilbert space, then all square roots ofT are related by unitary transformations. More precisely, ifT =A*A =B*B, then there exists aunitaryU such thatA =UB.
Indeed, takeB =T1/2 to be the unique non-negative square root ofT. IfT is strictly positive, thenB is invertible, and soU =AB−1 is unitary:
IfT is non-negative without being strictly positive, then the inverse ofB cannot be defined, but theMoore–Penrose pseudoinverseB+ can be. In that case, the operatorB+A is apartial isometry, that is, a unitary operator from the range ofT to itself. This can then be extended to a unitary operatorU on the whole space by setting it equal to the identity on the kernel ofT. More generally, this is true on an infinite-dimensional Hilbert space if, in addition,T hasclosed range. In general, ifA,B areclosed and densely defined operators on a Hilbert spaceH, andA* A =B* B, thenA = UB whereU is a partial isometry.
Square roots, and the unitary freedom of square roots, have applications throughoutfunctional analysis and linear algebra.
IfA is an invertible operator on a finite-dimensional Hilbert space, then there is a unique unitary operatorU and positive operatorP such that
this is the polar decomposition ofA. The positive operatorP is the unique positive square root of the positive operatorA∗A, andU is defined byU =AP−1.
IfA is not invertible, then it still has a polar composition in whichP is defined in the same way (and is unique). The unitary operatorU is not unique. Rather it is possible to determine a "natural" unitary operator as follows:AP+ is a unitary operator from the range ofA to itself, which can be extended by the identity on the kernel ofA∗. The resulting unitary operatorU then yields the polar decomposition ofA.
By Choi's result, a linear map
is completely positive if and only if it is of the form
wherek ≤nm. Let {Epq} ⊂Cn ×n be then2 elementary matrix units. The positive matrix
is called theChoi matrix of Φ. The Kraus operators correspond to the, not necessarily square, square roots ofMΦ: For any square rootB ofMΦ, one can obtain a family of Kraus operatorsVi by undoing the Vec operation to each columnbi ofB. Thus all sets of Kraus operators are related by partial isometries.
In quantum physics, a density matrix for ann-level quantum system is ann ×n complex matrixρ that is positive semidefinite with trace 1. Ifρ can be expressed as
where and Σpi = 1, the set
is said to be anensemble that describes the mixed stateρ. Notice {vi} is not required to be orthogonal. Different ensembles describing the stateρ are related by unitary operators, via the square roots ofρ. For instance, suppose
The trace 1 condition means
Let
andvi be the normalizedai. We see that
gives the mixed stateρ.