Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Rank (linear algebra)

From Wikipedia, the free encyclopedia
Dimension of the column space of a matrix

Inlinear algebra, therank of amatrixA is thedimension of thevector space generated (orspanned) by its columns.[1][2][3] This corresponds to the maximal number oflinearly independent columns ofA. This, in turn, is identical to the dimension of the vector space spanned by its rows.[4] Rank is thus a measure of the "nondegenerateness" of thesystem of linear equations andlinear transformation encoded byA. There are multiple equivalent definitions of rank. A matrix's rank is one of its most fundamental characteristics.

The rank is commonly denoted byrank(A) orrk(A);[2] sometimes the parentheses are not written, as inrankA.[i]

Main definitions

[edit]

In this section, we give some definitions of the rank of a matrix. Many definitions are possible; seeAlternative definitions for several of these.

Thecolumn rank ofA is thedimension of thecolumn space ofA, while therow rank ofA is the dimension of therow space ofA.

A fundamental result in linear algebra is that the column rank and the row rank are always equal. (Three proofs of this result are given in§ Proofs that column rank = row rank, below.) This number (i.e., the number of linearly independent rows or columns) is simply called therank ofA.

A matrix is said to havefull rank if its rank equals the largest possible for a matrix of the same dimensions, which is the lesser of the number of rows and columns. A matrix is said to berank-deficient if it does not have full rank. Therank deficiency of a matrix is the difference between the lesser of the number of rows and columns, and the rank.

The rank of alinear map or operatorΦ{\displaystyle \Phi } is defined as the dimension of itsimage:[5][6][7][8]rank(Φ):=dim(img(Φ)){\displaystyle \operatorname {rank} (\Phi ):=\dim(\operatorname {img} (\Phi ))}wheredim{\displaystyle \dim } is the dimension of a vector space, andimg{\displaystyle \operatorname {img} } is the image of a map.

Examples

[edit]

The matrix[101011011]{\displaystyle {\begin{bmatrix}1&0&1\\0&1&1\\0&1&1\end{bmatrix}}}has rank 2: the first two columns arelinearly independent, so the rank is at least 2, but since the third is a linear combination of the first two (the first column plus the second), the three columns are linearly dependent so the rank must be less than 3.

The matrixA=[11021102]{\displaystyle A={\begin{bmatrix}1&1&0&2\\-1&-1&0&-2\end{bmatrix}}}has rank 1: there are nonzero columns, so the rank is positive, but any pair of columns is linearly dependent. Similarly, thetransposeAT=[11110022]{\displaystyle A^{\mathrm {T} }={\begin{bmatrix}1&-1\\1&-1\\0&0\\2&-2\end{bmatrix}}}ofA has rank 1. Indeed, since the column vectors ofA are the row vectors of thetranspose ofA, the statement that the column rank of a matrix equals its row rank is equivalent to the statement that the rank of a matrix is equal to the rank of its transpose, i.e.,rank(A) = rank(AT).

Computing the rank of a matrix

[edit]

Rank from row echelon forms

[edit]
Main article:Gaussian elimination

A common approach to finding the rank of a matrix is to reduce it to a simpler form, generallyrow echelon form, byelementary row operations. Row operations do not change the row space (hence do not change the row rank), and, being invertible, map the column space to an isomorphic space (hence do not change the column rank). Once in row echelon form, the rank is clearly the same for both row rank and column rank, and equals the number ofpivots (or basic columns) and also the number of non-zero rows.

For example, the matrixA given byA=[121231350]{\displaystyle A={\begin{bmatrix}1&2&1\\-2&-3&1\\3&5&0\end{bmatrix}}}can be put in reduced row-echelon form by using the following elementary row operations:[121231350]2R1+R2R2[121013350]3R1+R3R3[121013013]R2+R3R3[121013000]2R2+R1R1[105013000] .{\displaystyle {\begin{aligned}{\begin{bmatrix}1&2&1\\-2&-3&1\\3&5&0\end{bmatrix}}&\xrightarrow {2R_{1}+R_{2}\to R_{2}} {\begin{bmatrix}1&2&1\\0&1&3\\3&5&0\end{bmatrix}}\xrightarrow {-3R_{1}+R_{3}\to R_{3}} {\begin{bmatrix}1&2&1\\0&1&3\\0&-1&-3\end{bmatrix}}\\&\xrightarrow {R_{2}+R_{3}\to R_{3}} \,\,{\begin{bmatrix}1&2&1\\0&1&3\\0&0&0\end{bmatrix}}\xrightarrow {-2R_{2}+R_{1}\to R_{1}} {\begin{bmatrix}1&0&-5\\0&1&3\\0&0&0\end{bmatrix}}~.\end{aligned}}}The final matrix (in reduced row echelon form) has two non-zero rows and thus the rank of matrixA is 2.

Computation

[edit]

When applied tofloating point computations on computers, basic Gaussian elimination (LU decomposition) can be unreliable, and a rank-revealing decomposition should be used instead. An effective alternative is thesingular value decomposition (SVD), but there are other less computationally expensive choices, such asQR decomposition with pivoting (so-calledrank-revealing QR factorization), which are still more numerically robust than Gaussian elimination. Numerical determination of rank requires a criterion for deciding when a value, such as a singular value from the SVD, should be treated as zero, a practical choice which depends on both the matrix and the application.

Proofs that column rank = row rank

[edit]

Proof using row reduction

[edit]

The fact that the column and row ranks of any matrix are equal is fundamental in linear algebra. Many proofs have been given. One of the most elementary ones has been sketched in§ Rank from row echelon forms. Here is a variant of this proof:

It is straightforward to show that neither the row rank nor the column rank are changed by anelementary row operation. AsGaussian elimination proceeds by elementary row operations, thereduced row echelon form of a matrix has the same row rank and the same column rank as the original matrix. Further elementary column operations allow putting the matrix in the form of anidentity matrix possibly bordered by rows and columns of zeros. Again, this changes neither the row rank nor the column rank. It is immediate that both the row and column ranks of this resulting matrix is the number of its nonzero entries.

We present two other proofs of this result. The first uses only basic properties oflinear combinations of vectors, and is valid over anyfield. The proof is based upon Wardlaw (2005).[9] The second usesorthogonality and is valid for matrices over thereal numbers; it is based upon Mackiw (1995).[4] Both proofs can be found in the book by Banerjee and Roy (2014).[10]

Proof using linear combinations

[edit]

LetA be anm ×n matrix. Let the column rank ofA ber, and letc1, ...,cr be any basis for the column space ofA. Place these as the columns of anm ×r matrixC. Every column ofA can be expressed as a linear combination of ther columns inC. This means that there is anr ×n matrixR such thatA =CR.R is the matrix whoseith column is formed from the coefficients giving theith column ofA as a linear combination of ther columns ofC. In other words,R is the matrix which contains the multiples for the bases of the column space ofA (which isC), which are then used to formA as a whole. Now, each row ofA is given by a linear combination of ther rows ofR. Therefore, the rows ofR form a spanning set of the row space ofA and, by theSteinitz exchange lemma, the row rank ofA cannot exceedr. This proves that the row rank ofA is less than or equal to the column rank ofA. This result can be applied to any matrix, so apply the result to the transpose ofA. Since the row rank of the transpose ofA is the column rank ofA and the column rank of the transpose ofA is the row rank ofA, this establishes the reverse inequality and we obtain the equality of the row rank and the column rank ofA. (Also seeRank factorization.)

Proof using orthogonality

[edit]

LetA be anm × n matrix with entries in thereal numbers whose row rank isr. Therefore, the dimension of the row space ofA isr. Letx1,x2, ...,xr be abasis of the row space ofA. We claim that the vectorsAx1,Ax2, ...,Axr arelinearly independent. To see why, consider a linear homogeneous relation involving these vectors with scalar coefficientsc1,c2, ...,cr:0=c1Ax1+c2Ax2++crAxr=A(c1x1+c2x2++crxr)=Av,{\displaystyle 0=c_{1}A\mathbf {x} _{1}+c_{2}A\mathbf {x} _{2}+\cdots +c_{r}A\mathbf {x} _{r}=A(c_{1}\mathbf {x} _{1}+c_{2}\mathbf {x} _{2}+\cdots +c_{r}\mathbf {x} _{r})=A\mathbf {v} ,}wherev =c1x1 +c2x2 + ⋯ +crxr. We make two observations: (a)v is a linear combination of vectors in the row space ofA, which implies thatv belongs to the row space ofA, and (b) sinceAv = 0, the vectorv isorthogonal to every row vector ofA and, hence, is orthogonal to every vector in the row space ofA. The facts (a) and (b) together imply thatv is orthogonal to itself, which proves thatv = 0 or, by the definition ofv,c1x1+c2x2++crxr=0.{\displaystyle c_{1}\mathbf {x} _{1}+c_{2}\mathbf {x} _{2}+\cdots +c_{r}\mathbf {x} _{r}=0.}But recall that thexi were chosen as a basis of the row space ofA and so are linearly independent. This implies thatc1 =c2 = ⋯ =cr = 0. It follows thatAx1,Ax2, ...,Axr are linearly independent.

EveryAxi is in the column space ofA. So,Ax1,Ax2, ...,Axr is a set ofr linearly independent vectors in the column space ofA and, hence, the dimension of the column space ofA (i.e., the column rank ofA) must be at least as big asr. This proves that row rank ofA is no larger than the column rank ofA. Now apply this result to the transpose ofA to get the reverse inequality and conclude as in the previous proof.

Alternative definitions

[edit]

In all the definitions in this section, the matrixA is taken to be anm ×n matrix over an arbitraryfieldF.

Dimension of image

[edit]

Given the matrixA{\displaystyle A}, there is an associatedlinear mappingf:FnFm{\displaystyle f:F^{n}\to F^{m}}defined byf(x)=Ax.{\displaystyle f(x)=Ax.}The rank ofA{\displaystyle A} is the dimension of the image off{\displaystyle f}. This definition has the advantage that it can be applied to any linear map without need for a specific matrix.

Rank in terms of nullity

[edit]

Given the same linear mappingf as above, the rank isn minus the dimension of thekernel off. Therank–nullity theorem states that this definition is equivalent to the preceding one.

Column rank – dimension of column space

[edit]

The rank ofA is the maximal number of linearly independent columnsc1,c2,,ck{\displaystyle \mathbf {c} _{1},\mathbf {c} _{2},\dots ,\mathbf {c} _{k}} ofA; this is thedimension of thecolumn space ofA (the column space being the subspace ofFm generated by the columns ofA, which is in fact just the image of the linear mapf associated toA).

Row rank – dimension of row space

[edit]

The rank ofA is the maximal number of linearly independent rows ofA; this is the dimension of therow space ofA.

Decomposition rank

[edit]

The rank ofA is the smallest positive integerk such thatA can be factored asA=CR{\displaystyle A=CR}, whereC is anm ×k matrix andR is ak ×n matrix. In fact, for all integersk, the following are equivalent:

  1. the column rank ofA is less than or equal tok,
  2. there existk columnsc1,,ck{\displaystyle \mathbf {c} _{1},\ldots ,\mathbf {c} _{k}} of sizem such that every column ofA is a linear combination ofc1,,ck{\displaystyle \mathbf {c} _{1},\ldots ,\mathbf {c} _{k}},
  3. there exist anm×k{\displaystyle m\times k} matrixC and ak×n{\displaystyle k\times n} matrixR such thatA=CR{\displaystyle A=CR} (whenk is the rank, this is arank factorization ofA),
  4. there existk rowsr1,,rk{\displaystyle \mathbf {r} _{1},\ldots ,\mathbf {r} _{k}} of sizen such that every row ofA is a linear combination ofr1,,rk{\displaystyle \mathbf {r} _{1},\ldots ,\mathbf {r} _{k}},
  5. the row rank ofA is less than or equal tok.

Indeed, the following equivalences are obvious:(1)(2)(3)(4)(5){\displaystyle (1)\Leftrightarrow (2)\Leftrightarrow (3)\Leftrightarrow (4)\Leftrightarrow (5)}.For example, to prove (3) from (2), takeC to be the matrix whose columns arec1,,ck{\displaystyle \mathbf {c} _{1},\ldots ,\mathbf {c} _{k}} from (2).To prove (2) from (3), takec1,,ck{\displaystyle \mathbf {c} _{1},\ldots ,\mathbf {c} _{k}} to be the columns ofC.

It follows from the equivalence(1)(5){\displaystyle (1)\Leftrightarrow (5)} that the row rank is equal to the column rank.

As in the case of the "dimension of image" characterization, this can be generalized to a definition of the rank of any linear map: the rank of a linear mapf :VW is the minimal dimensionk of an intermediate spaceX such thatf can be written as the composition of a mapVX and a mapXW. Unfortunately, this definition does not suggest an efficient manner to compute the rank (for which it is better to use one of the alternative definitions). Seerank factorization for details.

Rank in terms of singular values

[edit]

The rank ofA equals the number of non-zerosingular values, which is the same as the number of non-zero diagonal elements in Σ in thesingular value decompositionA=UΣV{\displaystyle A=U\Sigma V^{*}}.

Determinantal rank – size of largest non-vanishing minor

[edit]

The rank ofA is the largest order of any non-zerominor inA. (The order of a minor is the side-length of the square sub-matrix of which it is the determinant.) Like the decomposition rank characterization, this does not give an efficient way of computing the rank, but it is useful theoretically: a single non-zero minor witnesses a lower bound (namely its order) for the rank of the matrix, which can be useful (for example) to prove that certain operations do not lower the rank of a matrix.

A non-vanishingp-minor (p ×p submatrix with non-zero determinant) shows that the rows and columns of that submatrix are linearly independent, and thus those rows and columns of the full matrix are linearly independent (in the full matrix), so the row and column rank are at least as large as the determinantal rank; however, the converse is less straightforward. The equivalence of determinantal rank and column rank is a strengthening of the statement that if the span ofn vectors has dimensionp, thenp of those vectors span the space (equivalently, that one can choose a spanning set that is asubset of the vectors): the equivalence implies that a subset of the rows and a subset of the columns simultaneously define an invertible submatrix (equivalently, if the span ofn vectors has dimensionp, thenp of these vectors span the spaceand there is a set ofp coordinates on which they are linearly independent).

Tensor rank – minimum number of simple tensors

[edit]
Main articles:Tensor rank decomposition andTensor rank

The rank ofA is the smallest numberk such thatA can be written as a sum ofk rank 1 matrices, where a matrix is defined to have rank 1 if and only if it can be written as a nonzero productcr{\displaystyle c\cdot r} of a column vectorc and a row vectorr. This notion of rank is calledtensor rank; it can be generalized in theseparable models interpretation of thesingular value decomposition.

Properties

[edit]

We assume thatA is anm ×n matrix, and we define the linear mapf byf(x) =Ax as above.

Applications

[edit]

One useful application of calculating the rank of a matrix is the computation of the number of solutions of asystem of linear equations. According to theRouché–Capelli theorem, the system is inconsistent if the rank of theaugmented matrix is greater than the rank of thecoefficient matrix. If on the other hand, the ranks of these two matrices are equal, then the system must have at least one solution. The solution is unique if and only if the rank equals the number of variables. Otherwise the general solution hask free parameters wherek is the difference between the number of variables and the rank. In this case (and assuming the system of equations is in the real or complex numbers) the system of equations has infinitely many solutions.

Incontrol theory, the rank of a matrix can be used to determine whether alinear system iscontrollable, orobservable.

In the field ofcommunication complexity, the rank of the communication matrix of a function gives bounds on the amount of communication needed for two parties to compute the function.

Generalization

[edit]

There are different generalizations of the concept of rank to matrices over arbitraryrings, where column rank, row rank, dimension of column space, and dimension of row space of a matrix may be different from the others or may not exist.

Thinking of matrices astensors, thetensor rank generalizes to arbitrary tensors; for tensors of order greater than 2 (matrices are order 2 tensors), rank is very hard to compute, unlike for matrices.

There is a notion ofrank forsmooth maps betweensmooth manifolds. It is equal to the linear rank of thederivative.

Matrices as tensors

[edit]

Matrix rank should not be confused withtensor order, which is called tensor rank. Tensor order is the number of indices required to write atensor, and thus matrices all have tensor order 2. More precisely, matrices are tensors of type (1,1), having one row index and one column index, also called covariant order 1 and contravariant order 1; seeTensor (intrinsic definition) for details.

The tensor rank of a matrix can also mean the minimum number ofsimple tensors necessary to express the matrix as a linear combination, and that this definition does agree with matrix rank as here discussed.

See also

[edit]

Notes

[edit]
  1. ^Alternative notation includesρ(Φ){\displaystyle \rho (\Phi )} fromKatznelson & Katznelson (2008, p. 52, §2.5.1) andHalmos (1974, p. 90, § 50).
  2. ^Proof: Apply therank–nullity theorem to the inequalitydimker(AB)dimker(A)+dimker(B).{\displaystyle \dim \ker(AB)\leq \dim \ker(A)+\dim \ker(B).}
  3. ^Proof. The mapC:ker(ABC)/ker(BC)ker(AB)/ker(B){\displaystyle C:\ker(ABC)/\ker(BC)\to \ker(AB)/\ker(B)}is well-defined and injective. We thus obtain the inequality in terms of dimensions of kernel, which can then be converted to the inequality in terms of ranks by therank–nullity theorem.Alternatively, ifM{\displaystyle M} is a linear subspace thendim(AM)dim(M){\displaystyle \dim(AM)\leq \dim(M)}; apply this inequality to the subspace defined by the orthogonal complement of the image ofBC{\displaystyle BC} in the image ofB{\displaystyle B}, whose dimension isrank(B)rank(BC){\displaystyle \operatorname {rank} (B)-\operatorname {rank} (BC)}; its image underA{\displaystyle A} has dimensionrank(AB)rank(ABC){\displaystyle \operatorname {rank} (AB)-\operatorname {rank} (ABC)}.

References

[edit]
  1. ^Axler (2015) pp. 111-112, §§ 3.115, 3.119
  2. ^abRoman (2005) p. 48, § 1.16
  3. ^Bourbaki,Algebra, ch. II, §10.12, p. 359
  4. ^abMackiw, G. (1995), "A Note on the Equality of the Column and Row Rank of a Matrix",Mathematics Magazine,68 (4):285–286,doi:10.1080/0025570X.1995.11996337
  5. ^Hefferon (2020) p. 200, ch. 3, Definition 2.1
  6. ^Katznelson & Katznelson (2008) p. 52, § 2.5.1
  7. ^Valenza (1993) p. 71, § 4.3
  8. ^Halmos (1974) p. 90, § 50
  9. ^Wardlaw, William P. (2005), "Row Rank Equals Column Rank",Mathematics Magazine,78 (4):316–318,doi:10.1080/0025570X.2005.11953349,S2CID 218542661
  10. ^Banerjee, Sudipto; Roy, Anindya (2014),Linear Algebra and Matrix Analysis for Statistics, Texts in Statistical Science (1st ed.), Chapman and Hall/CRC,ISBN 978-1420095388
  11. ^Mirsky, Leonid (1955).An introduction to linear algebra. Dover Publications.ISBN 978-0-486-66434-7.{{cite book}}:ISBN / Date incompatibility (help)

Sources

[edit]

Further reading

[edit]
  • Roger A. Horn and Charles R. Johnson (1985).Matrix Analysis. Cambridge University Press.ISBN 978-0-521-38632-6.
  • Kaw, Autar K. Two Chapters from the book Introduction to Matrix Algebra: 1. Vectors[1] and System of Equations[2]
  • Mike Brookes: Matrix Reference Manual.[3]
Linear equations
Three dimensional Euclidean space
Matrices
Matrix decompositions
Relations and computations
Vector spaces
Structures
Multilinear algebra
Affine and projective
Numerical linear algebra
Retrieved from "https://en.wikipedia.org/w/index.php?title=Rank_(linear_algebra)&oldid=1313218615"
Category:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp