Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Hessenberg matrix

From Wikipedia, the free encyclopedia
Kind of square matrix in linear algebra

Inlinear algebra, aHessenberg matrix is a special kind ofsquare matrix, one that is "almost"triangular. To be exact, anupper Hessenberg matrix has zero entries below the firstsubdiagonal, and alower Hessenberg matrix has zero entries above the firstsuperdiagonal.[1] They are named afterKarl Hessenberg.[2]

AHessenberg decomposition is amatrix decomposition of a matrixA{\displaystyle A} into aunitary matrixP{\displaystyle P} and a Hessenberg matrixH{\displaystyle H} such thatPHP=A{\displaystyle PHP^{*}=A} whereP{\displaystyle P^{*}} denotes theconjugate transpose.

Definitions

[edit]

Upper Hessenberg matrix

[edit]

A squaren×n{\displaystyle n\times n} matrixA{\displaystyle A} is said to be inupper Hessenberg form or to be anupper Hessenberg matrix ifai,j=0{\displaystyle a_{i,j}=0} for alli,j{\displaystyle i,j} withi>j+1{\displaystyle i>j+1}.

An upper Hessenberg matrix is calledunreduced if all subdiagonal entries are nonzero, i.e. ifai+1,i0{\displaystyle a_{i+1,i}\neq 0} for alli{1,,n1}{\displaystyle i\in \{1,\ldots ,n-1\}}.[3]

Lower Hessenberg matrix

[edit]

A squaren×n{\displaystyle n\times n} matrixA{\displaystyle A} is said to be inlower Hessenberg form or to be alower Hessenberg matrix if its transpose{\displaystyle } is an upper Hessenberg matrix or equivalently ifai,j=0{\displaystyle a_{i,j}=0} for alli,j{\displaystyle i,j} withj>i+1{\displaystyle j>i+1}.

A lower Hessenberg matrix is calledunreduced if all superdiagonal entries are nonzero, i.e. ifai,i+10{\displaystyle a_{i,i+1}\neq 0} for alli{1,,n1}{\displaystyle i\in \{1,\ldots ,n-1\}}.

Examples

[edit]

Consider the following matrices.A=[1423341702340013]{\displaystyle A={\begin{bmatrix}1&4&2&3\\3&4&1&7\\0&2&3&4\\0&0&1&3\\\end{bmatrix}}}B=[1200523034375611]{\displaystyle B={\begin{bmatrix}1&2&0&0\\5&2&3&0\\3&4&3&7\\5&6&1&1\\\end{bmatrix}}}C=[1200520034375611]{\displaystyle C={\begin{bmatrix}1&2&0&0\\5&2&0&0\\3&4&3&7\\5&6&1&1\\\end{bmatrix}}}

The matrixA{\displaystyle A} is an upper unreduced Hessenberg matrix,B{\displaystyle B} is a lower unreduced Hessenberg matrix andC{\displaystyle C} is a lower Hessenberg matrix but is not unreduced.

Computer programming

[edit]

Many linear algebraalgorithms require significantly lesscomputational effort when applied totriangular matrices, and this improvement often carries over to Hessenberg matrices as well. If the constraints of a linear algebra problem do not allow a general matrix to be conveniently reduced to a triangular one, reduction to Hessenberg form is often the next best thing. In fact, reduction of any matrix to a Hessenberg form can be achieved in a finite number of steps (for example, throughHouseholder's transformation of unitary similarity transforms). Subsequent reduction of Hessenberg matrix to a triangular matrix can be achieved through iterative procedures, such as shiftedQR-factorization. Ineigenvalue algorithms, the Hessenberg matrix can be further reduced to a triangular matrix through Shifted QR-factorization combined with deflation steps. Reducing a general matrix to a Hessenberg matrix and then reducing further to a triangular matrix, instead of directly reducing a general matrix to a triangular matrix, often economizes the arithmetic involved in theQR algorithm for eigenvalue problems.

Reduction to Hessenberg matrix

[edit]

Householder transformations

[edit]

Anyn×n{\displaystyle n\times n} matrix can be transformed into a Hessenberg matrix by a similarity transformation usingHouseholder transformations. The following procedure for such a transformation is adapted fromA Second Course In Linear Algebra byGarcia & Horn.[4]

LetA{\displaystyle A} be any real or complexn×n{\displaystyle n\times n} matrix, then letA{\displaystyle A^{\prime }} be the(n1)×n{\displaystyle (n-1)\times n} submatrix ofA{\displaystyle A} constructed by removing the first row inA{\displaystyle A} and leta1{\displaystyle \mathbf {a} _{1}^{\prime }} be the first column ofA{\displaystyle A'}. Construct the(n1)×(n1){\displaystyle (n-1)\times (n-1)} householder matrixV1=I(n1)2www2{\displaystyle V_{1}=I_{(n-1)}-2{\frac {ww^{*}}{\|w\|^{2}}}} wherew={a12e1a1,a11=0a12e1+a11¯|a11|a1,a110{\displaystyle w={\begin{cases}\|\mathbf {a} _{1}^{\prime }\|_{2}\mathbf {e} _{1}-\mathbf {a} _{1}^{\prime }\;\;\;\;\;\;\;\;,\;\;\;a_{11}^{\prime }=0\\\|\mathbf {a} _{1}^{\prime }\|_{2}\mathbf {e} _{1}+{\frac {\overline {a_{11}^{\prime }}}{|a_{11}^{\prime }|}}\mathbf {a} _{1}^{\prime }\;\;\;,\;\;\;a_{11}^{\prime }\neq 0\\\end{cases}}}

This householder matrix will mapa1{\displaystyle \mathbf {a} _{1}^{\prime }} toa1e1{\displaystyle \|\mathbf {a} _{1}^{\prime }\|\mathbf {e} _{1}} and as such, the block matrixU1=[100V1]{\displaystyle U_{1}={\begin{bmatrix}1&\mathbf {0} \\\mathbf {0} &V_{1}\end{bmatrix}}} will map the matrixA{\displaystyle A} to the matrixU1A{\displaystyle U_{1}A} which has only zeros below the second entry of the first column. Now construct(n2)×(n2){\displaystyle (n-2)\times (n-2)} householder matrixV2{\displaystyle V_{2}} in a similar manner asV1{\displaystyle V_{1}} such thatV2{\displaystyle V_{2}} maps the first column ofA{\displaystyle A^{\prime \prime }} toa1e1{\displaystyle \|\mathbf {a} _{1}^{\prime \prime }\|\mathbf {e} _{1}}, whereA{\displaystyle A^{\prime \prime }} is the submatrix ofA{\displaystyle A^{\prime }} constructed by removing the first row and the first column ofA{\displaystyle A^{\prime }}, then letU2=[10001000V2]{\displaystyle U_{2}={\begin{bmatrix}1&0&0\\0&1&0\\0&0&V_{2}\end{bmatrix}}} which mapsU1A{\displaystyle U_{1}A} to the matrixU2U1A{\displaystyle U_{2}U_{1}A} which has only zeros below the first and second entry of the subdiagonal. Now constructV3{\displaystyle V_{3}} and thenU3{\displaystyle U_{3}} in a similar manner, but for the matrixA{\displaystyle A^{\prime \prime \prime }} constructed by removing the first row and first column ofA{\displaystyle A^{\prime \prime }} and proceed as in the previous steps. Continue like this for a total ofn2{\displaystyle n-2} steps.

By construction ofUk{\displaystyle U_{k}}, the firstk{\displaystyle k} columns of anyn×n{\displaystyle n\times n} matrix are invariant under multiplication byUk{\displaystyle U_{k}^{*}} from the right. Hence, any matrix can be transformed to an upper Hessenberg matrix by a similarity transformation of the formU(n2)((U2(U1AU1)U2))U(n2)=U(n2)U2U1A(U(n2)U2U1)=UAU{\displaystyle U_{(n-2)}(\dots (U_{2}(U_{1}AU_{1}^{*})U_{2}^{*})\dots )U_{(n-2)}^{*}=U_{(n-2)}\dots U_{2}U_{1}A(U_{(n-2)}\dots U_{2}U_{1})^{*}=UAU^{*}}.

Jacobi (Givens) rotations

[edit]

AJacobi rotation (also called Givens rotation) is an orthogonal matrix transformation in the form

AA=J(p,q,θ)TAJ(p,q,θ),{\displaystyle A\to A'=J(p,q,\theta )^{T}AJ(p,q,\theta )\;,}

whereJ(p,q,θ){\displaystyle J(p,q,\theta )},p<q{\displaystyle p<q}, is the Jacobi rotation matrix with all matrix elements equal zero except for

{J(p,q,θ)ii=1ip,qJ(p,q,θ)pp=cos(θ)J(p,q,θ)qq=cos(θ)J(p,q,θ)pq=sin(θ)J(p,q,θ)qp=sin(θ).{\displaystyle \left\{{\begin{aligned}J(p,q,\theta )_{ii}&{}=1\;\forall i\neq p,q\\J(p,q,\theta )_{pp}&{}=\cos(\theta )\\J(p,q,\theta )_{qq}&{}=\cos(\theta )\\J(p,q,\theta )_{pq}&{}=\sin(\theta )\\J(p,q,\theta )_{qp}&{}=-\sin(\theta )\;.\end{aligned}}\right.}

One can zero the matrix elementAp1,q{\displaystyle A'_{p-1,q}} by choosingthe rotation angleθ{\displaystyle \theta } to satisfy the equation

Ap1,psinθ+Ap1,qcosθ=0,{\displaystyle A_{p-1,p}\sin \theta +A_{p-1,q}\cos \theta =0\;,}

Now, the sequence of such Jacobi rotations with the following(p,q){\displaystyle (p,q)}

(p,q)=(2,3),(2,4),,(2,n),(3,4),,(3,n),,(n1,n){\displaystyle (p,q)=(2,3),(2,4),\dots ,(2,n),(3,4),\dots ,(3,n),\dots ,(n-1,n)}

reduces the matrixA{\displaystyle A} to the lower Hessenberg form.[5]

Properties

[edit]

Forn{1,2}{\displaystyle n\in \{1,2\}}, everyn×n{\displaystyle n\times n} matrix is both upper Hessenberg, and lower Hessenberg.[6]

The product of a Hessenberg matrix with a triangular matrix is again Hessenberg. More precisely, ifA{\displaystyle A} is upper Hessenberg andT{\displaystyle T} is upper triangular, thenAT{\displaystyle AT} andTA{\displaystyle TA} are upper Hessenberg.

A matrix that is both upper Hessenberg and lower Hessenberg is atridiagonal matrix, of which theJacobi matrix is an important example. This includes the symmetric or Hermitian Hessenberg matrices. A Hermitian matrix can be reduced to tri-diagonal real symmetric matrices.[7]

The Hessenberg matrix can be presented in a Jordan canonical form, withconfluent Vandermonde matrix as the similarity matrix (chapter 1.4.2 of[8]). For review of algorithms inverting this form of similarity matrix, necessary to construct that Jordan form, look into chapter 1.9 also of[8]. For spectral properties of Hessenberg, and matrix look also into chapter 1.9 of[8].

Hessenberg operator

[edit]

The Hessenberg operator is an infinite dimensional Hessenberg matrix. It commonly occurs as the generalization of theJacobi operator to a system oforthogonal polynomials for the space ofsquare-integrableholomorphic functions over some domain—that is, aBergman space. In this case, the Hessenberg operator is the right-shift operatorS{\displaystyle S}, given by[Sf](z)=zf(z).{\displaystyle [Sf](z)=zf(z).}

Theeigenvalues of each principal submatrix of the Hessenberg operator are given by thecharacteristic polynomial for that submatrix. These polynomials are called theBergman polynomials, and provide anorthogonal polynomial basis for Bergman space.

See also

[edit]

Notes

[edit]
  1. ^Horn & Johnson (1985), page 28;Stoer & Bulirsch (2002), page 251
  2. ^Biswa Nath Datta (2010) Numerical Linear Algebra and Applications, 2nd Ed., Society for Industrial and Applied Mathematics (SIAM)ISBN 978-0-89871-685-6, p. 307
  3. ^Horn & Johnson 1985, p. 35
  4. ^Ramon Garcia, Stephan; Horn, Roger (2017).A Second Course In Linear Algebra. Cambridge University Press.ISBN 9781107103818.
  5. ^Bini, Dario A.; Robol, Leonardo (2016). "Quasiseparable Hessenberg reduction of real diagonal plus low rank matrices and applications".Linear Algebra and Its Applications.502:186–213.arXiv:1501.07812.doi:10.1016/j.laa.2015.08.026.
  6. ^Lecture Notes. Notes for 2016-10-21 Cornell University
  7. ^"Computational Routines (eigenvalues) in LAPACK".sites.science.oregonstate.edu. Retrieved2020-05-24.
  8. ^abcMeurant, G. (2025).Hessenberg and Tridiagonal Matrices. SIAM.

References

[edit]

External links

[edit]
Matrix classes
Explicitly constrained entries
Constant
Conditions oneigenvalues or eigenvectors
Satisfying conditions onproducts orinverses
With specific applications
Used instatistics
Used ingraph theory
Used in science and engineering
Related terms
Retrieved from "https://en.wikipedia.org/w/index.php?title=Hessenberg_matrix&oldid=1335903002"
Category:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp