In mathematics, a Hermitian matrix (or self-adjoint matrix) is a complex square matrix that is equal to its own conjugate transpose—that is, the element in the -th row and -th column is equal to the complex conjugate of the element in the -th row and -th column, for all indices and :
or in matrix form:
Hermitian matrices can be understood as the complex extension of real symmetric matrices.
If the conjugate transpose of a matrix
A
AH,
Hermitian matrices are named after Charles Hermite, who demonstrated in 1855 that matrices of this form share a property with real symmetric matrices of always having real eigenvalues. Other, equivalent notations in common use are
AH=A\dagger=A\ast,
A\ast
Hermitian matrices can be characterized in a number of equivalent ways, some of which are listed below:
A square matrix
A
v,w,
\langle ⋅ , ⋅ \rangle
This is also the way that the more general concept of self-adjoint operator is defined.
An
n x {}n
A
A square matrix
A
Hermitian matrices are fundamental to quantum mechanics because they describe operators with necessarily real eigenvalues. An eigenvalue
a
\hat{A}
|\psi\rangle
In signal processing, Hermitian matrices are utilized in tasks like Fourier analysis and signal representation.[1] The eigenvalues and eigenvectors of Hermitian matrices play a crucial role in analyzing signals and extracting meaningful information.
Hermitian matrices are extensively studied in linear algebra and numerical analysis. They have well-defined spectral properties, and many numerical algorithms, such as the Lanczos algorithm, exploit these properties for efficient computations. Hermitian matrices also appear in techniques like singular value decomposition (SVD) and eigenvalue decomposition.
In statistics and machine learning, Hermitian matrices are used in covariance matrices, where they represent the relationships between different variables. The positive definiteness of a Hermitian covariance matrix ensures the well-definedness of multivariate distributions.[2]
Hermitian matrices are applied in the design and analysis of communications system, especially in the field of multiple-input multiple-output (MIMO) systems. Channel matrices in MIMO systems often exhibit Hermitian properties.
In graph theory, Hermitian matrices are used to study the spectra of graphs. The Hermitian Laplacian matrix is a key tool in this context, as it is used to analyze the spectra of mixed graphs.[3] The Hermitian-adjacency matrix of a mixed graph is another important concept, as it is a Hermitian matrix that plays a role in studying the energies of mixed graphs.[4]
In this section, the conjugate transpose of matrix
A
AH,
A
AT
A
\overline{A}.
See the following example:
The diagonal elements must be real, as they must be their own complex conjugate.
Well-known families of Hermitian matrices include the Pauli matrices, the Gell-Mann matrices and their generalizations. In theoretical physics such Hermitian matrices are often multiplied by imaginary coefficients,[5] [6] which results in skew-Hermitian matrices.
Here, we offer another useful Hermitian matrix using an abstract example. If a square matrix
A
A=BBH,
A
B
A
The entries on the main diagonal (top left to bottom right) of any Hermitian matrix are real.
Only the main diagonal entries are necessarily real; Hermitian matrices can have arbitrary complex-valued entries in their off-diagonal elements, as long as diagonally-opposite entries are complex conjugates.
A matrix that has only real entries is symmetric if and only if it is a Hermitian matrix. A real and symmetric matrix is simply a special case of a Hermitian matrix.
So, if a real anti-symmetric matrix is multiplied by a real multiple of the imaginary unit
i,
Every Hermitian matrix is a normal matrix. That is to say,
AAH=AHA.
The finite-dimensional spectral theorem says that any Hermitian matrix can be diagonalized by a unitary matrix, and that the resulting diagonal matrix has only real entries. This implies that all eigenvalues of a Hermitian matrix with dimension are real, and that has linearly independent eigenvectors. Moreover, a Hermitian matrix has orthogonal eigenvectors for distinct eigenvalues. Even if there are degenerate eigenvalues, it is always possible to find an orthogonal basis of consisting of eigenvectors of .
The sum of any two Hermitian matrices is Hermitian.
The inverse of an invertible Hermitian matrix is Hermitian as well.
The product of two Hermitian matrices and is Hermitian if and only if .
If A and B are Hermitian, then ABA is also Hermitian.
For an arbitrary complex valued vector the product
vHAv
vHAv=\left(vHAv\right)H.
The Hermitian complex -by- matrices do not form a vector space over the complex numbers,, since the identity matrix is Hermitian, but is not. However the complex Hermitian matrices do form a vector space over the real numbers . In the -dimensional vector space of complex matrices over, the complex Hermitian matrices form a subspace of dimension . If denotes the -by- matrix with a in the position and zeros elsewhere, a basis (orthonormal with respect to the Frobenius inner product) can be described as follows:
together with the set of matrices of the form
and the matrices
where
i
i=\sqrt{-1}~.
An example is that the four Pauli matrices form a complete basis for the vector space of all complex 2-by-2 Hermitian matrices over .
If orthonormal eigenvectors
u1,...,un
A=UΛUH
UUH=I=UHU
λj
Λ.
The singular values of
A
Since
A
A=UΛUH
U
A
A=U|Λ|sgn(Λ)UH
|Λ|
sgn(Λ)
|λ|
sgn(λ)
A
sgn(Λ)UH
UH
\pm1
|Λ|
A
The determinant of a Hermitian matrix is real:
(Alternatively, the determinant is the product of the matrix's eigenvalues, and as mentioned before, the eigenvalues of a Hermitian matrix are real.)
Additional facts related to Hermitian matrices include:
\left(A+AH\right)
\left(A-AH\right)
See main article: Rayleigh quotient.
In mathematics, for a given complex Hermitian matrix and nonzero vector, the Rayleigh quotient[9]
R(M,x),
For real matrices and vectors, the condition of being Hermitian reduces to that of being symmetric, and the conjugate transpose
xH
xT.
R(M,cx)=R(M,x)
c.
It can be shown that, for a given matrix, the Rayleigh quotient reaches its minimum value
λmin
x
vmin
R(M,x)\leqλmax
R(M,vmax)=λmax.
The Rayleigh quotient is used in the min-max theorem to get exact values of all eigenvalues. It is also used in eigenvalue algorithms to obtain an eigenvalue approximation from an eigenvector approximation. Specifically, this is the basis for Rayleigh quotient iteration.
The range of the Rayleigh quotient (for matrix that is not necessarily Hermitian) is called a numerical range (or spectrum in functional analysis). When the matrix is Hermitian, the numerical range is equal to the spectral norm. Still in functional analysis,
λmax