A
P
D
D
T:V\toV
V
T
T
A=PDP-1
P
D
T
Diagonalization is the process of finding the above
P
D
The geometric transformation represented by a diagonalizable matrix is an inhomogeneous dilation (or anisotropic scaling). That is, it can scale the space by a different amount in different directions. The direction of each eigenvector is scaled by a factor given by the corresponding eigenvalue.
A square matrix that is not diagonalizable is called defective. It can happen that a matrix
A
A=PDP-1
P
D
A
Many results for diagonalizable matrices hold only over an algebraically closed field (such as the complex numbers). In this case, diagonalizable matrices are dense in the space of all matrices, which means any defective matrix can be deformed into a diagonalizable matrix by a small perturbation; and the Jordan–Chevalley decomposition states that any matrix is uniquely the sum of a diagonalizable matrix and a nilpotent matrix. Over an algebraically closed field, diagonalizable matrices are equivalent to semi-simple matrices.
A square
n x n
A
F
n x n
P
P-1AP
The fundamental fact about diagonalizable maps and matrices is expressed by the following:
n x n
A
F
n
Fn
A
P
P-1AP
A
P
A
T:V\toV
V
T
T
The following sufficient (but not necessary) condition is often useful.
n x n
A
F
n
n
P
A
A
T:V\toV
n=\dim(V)
n
n
F
Let
A
A
A
F
An
n
A
An
A
Over the complex numbers
\Complex
n x n
The Jordan–Chevalley decomposition expresses an operator as the sum of its semisimple (i.e., diagonalizable) part and its nilpotent part. Hence, a matrix is diagonalizable if and only if its nilpotent part is zero. Put in another way, a matrix is diagonalizable if each block in its Jordan form has no nilpotent part; i.e., each "block" is a one-by-one matrix.
Consider the two following arbitrary bases
E=\{{\boldsymbol{e}i|\foralli\in[n]}\}
F=\{{\boldsymbol{\alpha}i|\foralli\in[n]}\}
AE
AE\boldsymbol{\alpha}E,i=λi\boldsymbol{\alpha}E,i
The alpha eigenvectors are written also with respect to the E basis. Since the set F is both a set of eigenvectors for matrix A and it spans some arbitrary vector space, then we say that there exists a matrix
DF
AE
AE
S
DF=
F | |
S | |
E |
AE
-1F | |
S | |
E |
where
F | |
S | |
E |
P
-1F | |
S | |
E |
=
E | |
P | |
F |
Both
S
P
AE
A
If a matrix
A
P-1AP=\begin{bmatrix} λ1&0& … &0\\ 0&λ2& … &0\\ \vdots&\vdots&\ddots&\vdots\\ 0&0& … &λn \end{bmatrix}=D,
then:
AP=P\begin{bmatrix} λ1&0& … &0\\ 0&λ2& … &0\\ \vdots&\vdots&\ddots&\vdots\\ 0&0& … &λn \end{bmatrix}.
The transition matrix S has the E-basis vectors as columns written in the basis F. Inversely, the inverse transition matrix P has F-basis vectors
\boldsymbol{\alpha}i
P=\begin{bmatrix}\boldsymbol{\alpha}E,1&\boldsymbol{\alpha}E,2& … &\boldsymbol{\alpha}E,n\end{bmatrix},
as a result we can write:
In block matrix form, we can consider the A-matrix to be a matrix of 1x1 dimensions whilst P is a 1xn dimensional matrix. The D-matrix can be written in full form with all the diagonal elements as an nxn dimensional matrix:
A\begin{bmatrix}\boldsymbol{\alpha}E,1&\boldsymbol{\alpha}E,2& … &\boldsymbol{\alpha}E,n\end{bmatrix}=\begin{bmatrix}\boldsymbol{\alpha}E,1&\boldsymbol{\alpha}E,2& … &\boldsymbol{\alpha}E,n\end{bmatrix} \begin{bmatrix} λ1&0& … &0\\ 0&λ2& … &0\\ \vdots&\vdots&\ddots&\vdots\\ 0&0& … &λn \end{bmatrix}.
Performing the above matrix multiplication we end up with the following result:
\endTaking each component of the block matrix individually on both sides, we end up with the following:
A\boldsymbol{\alpha}i=λi\boldsymbol{\alpha}i (i=1,2,...,n).
So the column vectors of
P
P
P-1
When a complex matrix
A\inCn x
A
P
A\inRn x
Rn
P
For most practical work matrices are diagonalized numerically using computer software. Many algorithms exist to accomplish this.
See also: Weight (representation theory).
A set of matrices is said to be simultaneously diagonalizable if there exists a single invertible matrix
P
P-1AP
A
The set of all
n x n
n>1
\begin{bmatrix}1&0\ 0&0\end{bmatrix} and \begin{bmatrix}1&1\ 0&0\end{bmatrix}
are diagonalizable but not simultaneously diagonalizable because they do not commute.
A set consists of commuting normal matrices if and only if it is simultaneously diagonalizable by a unitary matrix; that is, there exists a unitary matrix
U
U*AU
A
In the language of Lie theory, a set of simultaneously diagonalizable matrices generates a toral Lie algebra.
C
QTAQ
AAT=ATA
In general, a rotation matrix is not diagonalizable over the reals, but all rotation matrices are diagonalizable over the complex field. Even if a matrix is not diagonalizable, it is always possible to "do the best one can", and find a matrix with the same properties consisting of eigenvalues on the leading diagonal, and either ones or zeroes on the superdiagonal – known as Jordan normal form.
Some matrices are not diagonalizable over any field, most notably nonzero nilpotent matrices. This happens more generally if the algebraic and geometric multiplicities of an eigenvalue do not coincide. For instance, consider
C=\begin{bmatrix}0&1\ 0&0\end{bmatrix}.
This matrix is not diagonalizable: there is no matrix
U
U-1CU
C
Some real matrices are not diagonalizable over the reals. Consider for instance the matrix
B=\left[\begin{array}{rr}0&1\ -1&0\end{array}\right].
The matrix
B
Q
Q-1BQ
B
Q=\begin{bmatrix}1&i\ i&1\end{bmatrix},
then
Q-1BQ
B
Note that the above examples show that the sum of diagonalizable matrices need not be diagonalizable.
Diagonalizing a matrix is the same process as finding its eigenvalues and eigenvectors, in the case that the eigenvectors form a basis. For example, consider the matrix
A=\left[\begin{array}{rrr} 0&1&-2\\ 0&1&0\\ 1&-1&3 \end{array}\right].
p(λ)=\det(λI-A)
\left(I-A\right)v=0
v1=(1,1,0)
\left(2I-A\right)v=0
Avi=λivi
P
P
P-1AP
Note that there is no preferred order of the eigenvectors in changing the order of the eigenvectors in
P
Diagonalization can be used to efficiently compute the powers of a matrix
\begin{align}Ak&=\left(PDP-1\right)k=\left(PDP-1\right)\left(PDP-1\right) … \left(PDP-1\right)\\ &=PD\left(P-1P\right)D\left(P-1P\right) … \left(P-1P\right)DP-1=PDkP-1, \end{align}
and the latter is easy to calculate since it only involves the powers of a diagonal matrix. For example, for the matrix
A
λ=1,1,2
\begin{align} Ak=PDkP-1&=\left[\begin{array}{rrr} 1&0&1\\ 1&2&0\\ 0&1&-1 \end{array}\right] \begin{bmatrix}1k&0&0\ 0&1k&0\ 0&0&2k\end{bmatrix} \left[\begin{array}{rrr} 1&0&1\\ 1&2&0\\ 0&1&-1 \end{array}\right]-1\\[1em] &=\begin{bmatrix} 2-2k&-1+2k&2-2k\\ 0&1&0\\ -1+2k&1-2k&-1+2k\end{bmatrix}. \end{align}
This approach can be generalized to matrix exponential and other matrix functions that can be defined as power series. For example, defining we have:
\begin{align} \exp(A)=P\exp(D)P-1&=\left[\begin{array}{rrr} 1&0&1\\ 1&2&0\\ 0&1&-1 \end{array}\right] \begin{bmatrix}e1&0&0\ 0&e1&0\ 0&0&e2\end{bmatrix} \left[\begin{array}{rrr} 1&0&1\\ 1&2&0\\ 0&1&-1 \end{array}\right]-1\\[1em] &=\begin{bmatrix} 2e-e2&-e+e2&2e-2e2\\ 0&e&0\\ -e+e2&e-e2&-e+2e2 \end{bmatrix}. \end{align}
This is particularly useful in finding closed form expressions for terms of linear recursive sequences, such as the Fibonacci numbers.
For example, consider the following matrix:
M=\begin{bmatrix}a&b-a\ 0&b\end{bmatrix}.
Calculating the various powers of
M
M2=\begin{bmatrix}a2&b2-a2\ 0&b2\end{bmatrix}, M3=\begin{bmatrix}a3&b3-a3\ 0&b3\end{bmatrix}, M4=\begin{bmatrix}a4&b4-a4\ 0&b4\end{bmatrix}, \ldots
The above phenomenon can be explained by diagonalizing To accomplish this, we need a basis of
\R2
u=\begin{bmatrix}1\ 0\end{bmatrix}=e1, v=\begin{bmatrix}1\ 1\end{bmatrix}=e1+e2,
where ei denotes the standard basis of Rn. The reverse change of basis is given by
e1=u, e2=v-u.
Straightforward calculations show that
Mu=au, Mv=bv.
Thus, a and b are the eigenvalues corresponding to u and v, respectively. By linearity of matrix multiplication, we have that
Mnu=anu, Mnv=bnv.
Switching back to the standard basis, we have
\begin{align} Mne1&=Mnu=ane1,\\ Mne2&=Mn\left(v-u\right)=bnv-anu=\left(bn-an\right)e1+
ne | |
b | |
2. \end{align} |
The preceding relations, expressed in matrix form, are
Mn=\begin{bmatrix}an&bn-an\ 0&bn\end{bmatrix},
thereby explaining the above phenomenon.
In quantum mechanical and quantum chemical computations matrix diagonalization is one of the most frequently applied numerical processes. The basic reason is that the time-independent Schrödinger equation is an eigenvalue equation, albeit in most of the physical situations on an infinite dimensional Hilbert space.
A very common approximation is to truncate Hilbert space to finite dimension, after which the Schrödinger equation can be formulated as an eigenvalue problem of a real symmetric, or complex Hermitian matrix. Formally this approximation is founded on the variational principle, valid for Hamiltonians that are bounded from below.
First-order perturbation theory also leads to matrix eigenvalue problem for degenerate states.