Matrix decomposition explained

In the mathematical discipline of linear algebra, a matrix decomposition or matrix factorization is a factorization of a matrix into a product of matrices. There are many different matrix decompositions; each finds use among a particular class of problems.

Example

In numerical analysis, different decompositions are used to implement efficient matrix algorithms.

Ax=b

, the matrix A can be decomposed via the LU decomposition. The LU decomposition factorizes a matrix into a lower triangular matrix L and an upper triangular matrix U. The systems

L(Ux)=b

and

Ux=L-1b

require fewer additions and multiplications to solve, compared with the original system

Ax=b

, though one might require significantly more digits in inexact arithmetic such as floating point.

Similarly, the QR decomposition expresses A as QR with Q an orthogonal matrix and R an upper triangular matrix. The system Q(Rx) = b is solved by Rx = QTb = c, and the system Rx = c is solved by 'back substitution'. The number of additions and multiplications required is about twice that of using the LU solver, but no more digits are required in inexact arithmetic because the QR decomposition is numerically stable.

Decompositions related to solving systems of linear equations

LU decomposition

See main article: LU decomposition.

A=LU

, where L is lower triangular and U is upper triangular.

A=LDU

, where L is lower triangular with ones on the diagonal, U is upper triangular with ones on the diagonal, and D is a diagonal matrix.

PA=LU

, where L is lower triangular, U is upper triangular, and P is a permutation matrix.

Ax=b

. These decompositions summarize the process of Gaussian elimination in matrix form. Matrix P represents any row interchanges carried out in the process of Gaussian elimination. If Gaussian elimination produces the row echelon form without requiring any row interchanges, then P = I, so an LU decomposition exists.

LU reduction

See main article: LU reduction.

Block LU decomposition

See main article: Block LU decomposition.

Rank factorization

See main article: Rank factorization.

A=CF

where C is an m-by-r full column rank matrix and F is an r-by-n full row rank matrix

Ax=b

.

Cholesky decomposition

See main article: Cholesky decomposition.

A

A=U*U

, where

U

is upper triangular with real positive diagonal entries

A

is Hermitian and positive semi-definite, then it has a decomposition of the form

A=U*U

if the diagonal entries of

U

are allowed to be zero

A

is real and symmetric,

U

has all real elements

QR decomposition

See main article: QR decomposition.

A=QR

where

Q

is a unitary matrix of size m-by-m, and

R

is an upper triangular matrix of size m-by-n

A

is of full rank, then there exists a single

R

that has all positive diagonal elements. If

A

is square, also

Q

is unique.

Ax=b

. The fact that

Q

is orthogonal means that

QTQ=I

, so that

Ax=b

is equivalent to

Rx=QTb

, which is very easy to solve since

R

is triangular.

RRQR factorization

See main article: RRQR factorization.

Interpolative decomposition

See main article: Interpolative decomposition.

Decompositions based on eigenvalues and related concepts

Eigendecomposition

See main article: Eigendecomposition (matrix).

A=VDV-1

, where D is a diagonal matrix formed from the eigenvalues of A, and the columns of V are the corresponding eigenvectors of A.

AV=VD

.

V

is invertible if and only if the n eigenvectors are linearly independent (that is, each eigenvalue has geometric multiplicity equal to its algebraic multiplicity). A sufficient (but not necessary) condition for this to happen is that all the eigenvalues are different (in this case geometric and algebraic multiplicity are equal to 1)

AA*=A*A

, where

A*

is a conjugate transpose) can be eigendecomposed. For a normal matrix A (and only for a normal matrix), the eigenvectors can also be made orthonormal (

VV*=I

) and the eigendecomposition reads as

A=VDV*

. In particular all unitary, Hermitian, or skew-Hermitian (in the real-valued case, all orthogonal, symmetric, or skew-symmetric, respectively) matrices are normal and therefore possess this property.

A=VDVT

, where both D and V are real-valued.

xt+1=Axt

starting from the initial condition

x0=c

is solved by

xt=Atc

, which is equivalent to

xt=VDtV-1c

, where V and D are the matrices formed from the eigenvectors and eigenvalues of A. Since D is diagonal, raising it to power

Dt

, just involves raising each element on the diagonal to the power t. This is much easier to do and understand than raising A to power t, since A is usually not diagonal.

Jordan decomposition

The Jordan normal form and the Jordan–Chevalley decomposition

Schur decomposition

See main article: Schur decomposition.

A=UTU*

, where U is a unitary matrix,

U*

is the conjugate transpose of U, and T is an upper triangular matrix called the complex Schur form which has the eigenvalues of A along its diagonal.

Real Schur decomposition

V

and

S

only contain real numbers. One can always write

A=VSVT

where V is a real orthogonal matrix,

VT

is the transpose of V, and S is a block upper triangular matrix called the real Schur form. The blocks on the diagonal of S are of size 1×1 (in which case they represent real eigenvalues) or 2×2 (in which case they are derived from complex conjugate eigenvalue pairs).

QZ decomposition

A=QSZ*

and

B=QTZ*

where Q and Z are unitary matrices, the * superscript represents conjugate transpose, and S and T are upper triangular matrices.

λi=Sii/Tii

, are the generalized eigenvalues that solve the generalized eigenvalue problem

Av=λBv

(where

λ

is an unknown scalar and v is an unknown nonzero vector).

A=QSZT

and

B=QTZT

where A, B, Q, Z, S, and T are matrices containing real numbers only. In this case Q and Z are orthogonal matrices, the T superscript represents transposition, and S and T are block upper triangular matrices. The blocks on the diagonal of S and T are of size 1×1 or 2×2.

Takagi's factorization

A=VDVT

, where D is a real nonnegative diagonal matrix, and V is unitary.

VT

denotes the matrix transpose of V.

AA*=VD2V-1

.

V-1

instead of

VT

. Moreover, if A is not real, it is not Hermitian and the form using

V*

also does not apply.

Singular value decomposition

See main article: Singular value decomposition.

A=UDV*

, where D is a nonnegative diagonal matrix, and U and V satisfy

U*U=I,V*V=I

. Here

V*

is the conjugate transpose of V (or simply the transpose, if V contains real numbers only), and I denotes the identity matrix (of some dimension).

A

are always uniquely determined.

U

and

V

need not to be unique in general.

Scale-invariant decompositions

Refers to variants of existing matrix decompositions, such as the SVD, that are invariant with respect to diagonal scaling.

A=DUSV*E

, where S is a unique nonnegative diagonal matrix of scale-invariant singular values, U and V are unitary matrices,

V*

is the conjugate transpose of V, and positive diagonal matrices D and E.

A

(given by the diagonal elements of S) are always uniquely determined. Diagonal matrices D and E, and unitary U and V, are not necessarily unique in general.

Analogous scale-invariant decompositions can be derived from other matrix decompositions; for example, to obtain scale-invariant eigenvalues.

Hessenberg decomposition

A=PHP*

where

H

is the Hessenberg matrix and

P

is a unitary matrix.

Complete orthogonal decomposition

See main article: Complete orthogonal decomposition.

A=UTV*

, where T is a triangular matrix, and U and V are unitary matrices.

Other decompositions

Polar decomposition

See main article: Polar decomposition.

A=UP

(right polar decomposition) or

A=P'U

(left polar decomposition), where U is a unitary matrix and P and P are positive semidefinite Hermitian matrices.

P

is always unique and equal to

\sqrt{A*A}

(which is always hermitian and positive semidefinite). If

A

is invertible, then

U

is unique.

P

can be written as

P=VDV*

. Since

P

is positive semidefinite, all elements in

D

are non-negative. Since the product of two unitary matrices is unitary, taking

W=UV

one can write

A=U(VDV*)=WDV*

which is the singular value decomposition. Hence, the existence of the polar decomposition is equivalent to the existence of the singular value decomposition.

Algebraic polar decomposition

A=QS

, where Q is a complex orthogonal matrix and S is complex symmetric matrix.

ATA

has no negative real eigenvalues, then the decomposition is unique.[4]

AAT

being similar to

ATA

.

A=RC

, where R is a real matrix and C is a circular matrix.

Mostow's decomposition

A=UeiMeS

, where U is unitary, M is real anti-symmetric and S is real symmetric.
S2
A=U
2e
iM2
e
, where U2 is unitary, M2 is real anti-symmetric and S2 is real symmetric.

Sinkhorn normal form

See main article: Sinkhorn's theorem.

A=D1SD2

, where S is doubly stochastic and D1 and D2 are real diagonal matrices with strictly positive elements.

Sectoral decomposition

S\alpha=\left\{rei\inC\midr>0,|\theta|\le\alpha<

\pi
2

\right\}

.

A=CZC*

, where C is an invertible complex matrix and

Z=

i\theta1
\operatorname{diag}\left(e
i\thetan
,\ldots,e

\right)

with all

\left|\thetaj\right|\le\alpha

.[6] [7]

Williamson's normal form

A=ST\operatorname{diag}(D,D)S

, where

S\inSp(2n)

is a symplectic matrix and D is a nonnegative n-by-n diagonal matrix.[8]

Matrix square root

See main article: Square root of a matrix.

A=BB

, not unique in general.

A

, there is a unique positive semidefinite

B

such that

A=B*B=BB

.

Generalizations

There exist analogues of the SVD, QR, LU and Cholesky factorizations for quasimatrices and cmatrices or continuous matrices. A ‘quasimatrix’ is, like a matrix, a rectangular scheme whose elements are indexed, but one discrete index is replaced by a continuous index. Likewise, a ‘cmatrix’, is continuous in both indices. As an example of a cmatrix, one can think of the kernel of an integral operator.

These factorizations are based on early work by, and . For an account, and a translation to English of the seminal papers, see .

See also

References

Bibliography

External links

Notes and References

  1. Book: Lay, David C.. Linear algebra and its applications. 2016. Steven R. Lay, Judith McDonald. 978-1-292-09223-2. Fifth Global. Harlow. 142. 920463015.
  2. If a non-square matrix is used, however, then the matrix U will also have the same rectangular shape as the original matrix A. And so, calling the matrix U upper triangular would be incorrect as the correct term would be that U is the 'row echelon form' of A. Other than this, there are no differences in LU factorization for square and non-square matrices.
  3. Piziak. R.. Odell. P. L.. Full Rank Factorization of Matrices. Mathematics Magazine. 1 June 1999. 72. 3. 193. 10.2307/2690882. 2690882.
  4. Bhatia. Rajendra. 2013-11-15. The bipolar decomposition. Linear Algebra and Its Applications. 439. 10. 3031–3037. 10.1016/j.laa.2013.09.006.
  5. Book: Matrix Information Geometry. Nielsen. Frank. Bhatia. Rajendra. Springer. 2012. 9783642302329. 224. en. 10.1007/978-3-642-30232-9. 1007.4402. 118466496 .
  6. Zhang. Fuzhen. A matrix decomposition and its applications. Linear and Multilinear Algebra. 63. 10. 30 June 2014. 2033–2042. 10.1080/03081087.2014.933219. 19437967 .
  7. Drury. S.W.. Fischer determinantal inequalities and Highamʼs Conjecture. Linear Algebra and Its Applications. November 2013. 439. 10. 3129–3133. 10.1016/j.laa.2013.08.031. free.
  8. Idel. Martin. Soto Gaona. Sebastián. Wolf. Michael M.. 2017-07-15. Perturbation bounds for Williamson's symplectic normal form. Linear Algebra and Its Applications. 525. 45–58. 10.1016/j.laa.2017.03.013. 1609.01338. 119578994 .