Matrix exponential explained

In mathematics, the matrix exponential is a matrix function on square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exponential gives the exponential map between a matrix Lie algebra and the corresponding Lie group.

Let be an real or complex matrix. The exponential of, denoted by or, is the matrix given by the power series

e^X = \sum_^\infty \frac X^k

where

X0

is defined to be the identity matrix

I

with the same dimensions as

X

.[1] The series always converges, so the exponential of is well-defined.

Equivalently,e^X = \lim_ \left(I + \frac \right)^k

where is the identity matrix.

When is an diagonal matrix then will be an diagonal matrix with each diagonal element equal to the ordinary exponential applied to the corresponding diagonal element of .

Properties

Elementary properties

Let and be complex matrices and let and be arbitrary complex numbers. We denote the identity matrix by and the zero matrix by 0. The matrix exponential satisfies the following properties.[2]

We begin with the properties that are immediate consequences of the definition as a power series:

The next key result is this one:

XY=YX

then

eXeY=eX+Y

.

The proof of this identity is the same as the standard power-series argument for the corresponding identity for the exponential of real numbers. That is to say, as long as

X

and

Y

commute, it makes no difference to the argument whether

X

and

Y

are numbers or matrices. It is important to note that this identity typically does not hold if

X

and

Y

do not commute (see Golden-Thompson inequality below).

Consequences of the preceding identity are the following:

Using the above results, we can easily verify the following claims. If is symmetric then is also symmetric, and if is skew-symmetric then is orthogonal. If is Hermitian then is also Hermitian, and if is skew-Hermitian then is unitary.

Finally, a Laplace transform of matrix exponentials amounts to the resolvent,\int_0^\infty e^e^\,dt = (sI - X)^for all sufficiently large positive values of .

Linear differential equation systems

See main article: Matrix differential equation.

One of the reasons for the importance of the matrix exponential is that it can be used to solve systems of linear ordinary differential equations. The solution of \frac y(t) = Ay(t), \quad y(0) = y_0, where is a constant matrix and y is a column vector, is given by y(t) = e^ y_0.

The matrix exponential can also be used to solve the inhomogeneous equation \frac y(t) = Ay(t) + z(t), \quad y(0) = y_0. See the section on applications below for examples.

There is no closed-form solution for differential equations of the form \frac y(t) = A(t) \, y(t), \quad y(0) = y_0, where is not constant, but the Magnus series gives the solution as an infinite sum.

The determinant of the matrix exponential

By Jacobi's formula, for any complex square matrix the following trace identity holds:[3]

In addition to providing a computational tool, this formula demonstrates that a matrix exponential is always an invertible matrix. This follows from the fact that the right hand side of the above equation is always non-zero, and so, which implies that must be invertible.

In the real-valued case, the formula also exhibits the map\exp \colon M_n(\R) \to \mathrm(n, \R)to not be surjective, in contrast to the complex case mentioned earlier. This follows from the fact that, for real-valued matrices, the right-hand side of the formula is always positive, while there exist invertible matrices with a negative determinant.

Real symmetric matrices

The matrix exponential of a real symmetric matrix is positive definite. Let

S

be an real symmetric matrix and

x\in\Rn

a column vector. Using the elementary properties of the matrix exponential and of symmetric matrices, we have:

x^Te^Sx=x^Te^e^x=x^T(e^)^Te^x =(e^x)^Te^x=\lVert e^x\rVert^2\geq 0.

Since

eS/2

is invertible, the equality only holds for

x=0

, and we have

xTeSx>0

for all non-zero

x

. Hence

eS

is positive definite.

The exponential of sums

For any real numbers (scalars) and we know that the exponential function satisfies . The same is true for commuting matrices. If matrices and commute (meaning that), then,e^ = e^Xe^Y.

However, for matrices that do not commute the above equality does not necessarily hold.

The Lie product formula

Even if and do not commute, the exponential can be computed by the Lie product formula[4] e^ = \lim_ \left(e^e^\right)^k.

Using a large finite to approximate the above is basis of the Suzuki-Trotter expansion, often used in numerical time evolution.

The Baker–Campbell–Hausdorff formula

In the other direction, if and are sufficiently small (but not necessarily commuting) matrices, we havee^Xe^Y = e^Z,where may be computed as a series in commutators of and by means of the Baker–Campbell–Hausdorff formula:[5] Z = X + Y + \frac[X,Y] + \frac[X,[X,Y]] - \frac[Y,[X,Y]]+ \cdots,where the remaining terms are all iterated commutators involving and . If and commute, then all the commutators are zero and we have simply .

Inequalities for exponentials of Hermitian matrices

See main article: Golden–Thompson inequality. For Hermitian matrices there is a notable theorem related to the trace of matrix exponentials.

If and are Hermitian matrices, then[6] \operatorname\exp(A + B) \leq \operatorname\left[\exp(A)\exp(B)\right].

There is no requirement of commutativity. There are counterexamples to show that the Golden–Thompson inequality cannot be extended to three matrices – and, in any event, is not guaranteed to be real for Hermitian,, . However, Lieb proved[7] [8] that it can be generalized to three matrices if we modify the expression as follows\operatorname\exp(A + B + C) \leq \int_0^\infty \mathrmt\, \operatorname\left[e^A\left(e^{-B} + t\right)^{-1}e^C \left(e^{-B} + t\right)^{-1}\right].

The exponential map

The exponential of a matrix is always an invertible matrix. The inverse matrix of is given by . This is analogous to the fact that the exponential of a complex number is always nonzero. The matrix exponential then gives us a map\exp \colon M_n(\Complex) \to \mathrm(n, \Complex)from the space of all n×n matrices to the general linear group of degree, i.e. the group of all n×n invertible matrices. In fact, this map is surjective which means that every invertible matrix can be written as the exponential of some other matrix[9] (for this, it is essential to consider the field C of complex numbers and not R).

For any two matrices and,\left\| e^ - e^X\right\| \le \|Y\| e^ e^,

where denotes an arbitrary matrix norm. It follows that the exponential map is continuous and Lipschitz continuous on compact subsets of .

The mapt \mapsto e^, \qquad t \in \Rdefines a smooth curve in the general linear group which passes through the identity element at .

In fact, this gives a one-parameter subgroup of the general linear group sincee^e^ = e^.

The derivative of this curve (or tangent vector) at a point t is given byThe derivative at is just the matrix X, which is to say that X generates this one-parameter subgroup.

More generally,[10] for a generic -dependent exponent,,

Taking the above expression outside the integral sign and expanding the integrand with the help of the Hadamard lemma one can obtain the following useful expression for the derivative of the matrix exponent,[11] \left(\frace^\right)e^ = \fracX(t) + \frac \left[X(t), \frac{d}{dt}X(t)\right] + \frac \left[X(t), \left[X(t), \frac{d}{dt}X(t)\right]\right] + \cdots

The coefficients in the expression above are different from what appears in the exponential. For a closed form, see derivative of the exponential map.

Directional derivatives when restricted to Hermitian matrices

Let

X

be a

n x n

Hermitian matrix with distinct eigenvalues. Let

X=Erm{diag}(Λ)E*

be its eigen-decomposition where

E

is a unitary matrix whose columns are the eigenvectors of

X

,

E*

is its conjugate transpose, and

Λ=\left(λ1,\ldots,λn\right)

the vector of corresponding eigenvalues. Then, for any

n x n

Hermitian matrix

V

, the directional derivative of

\exp:X\toeX

at

X

in the direction

V

is[12] [13] D \exp (X) [V]\triangleq\lim_ \frac\left(\displaystylee^-e^\right)=E(G \odot \bar) E^*where

\bar{V}=E*VE

, the operator

\odot

denotes the Hadamard product, and, for all

1\leqi,j\leqn

, the matrix

G

is defined asG_ = \left\

Notes and References

  1. Equation 2.1
  2. Proposition 2.3
  3. Theorem 2.12
  4. Theorem 2.11
  5. Chapter 5
  6. Book: Bhatia, R. . Matrix Analysis . Graduate Texts in Mathematics. 978-0-387-94846-1 . 1997 . Springer . 169.
  7. 10.1016/0001-8708(73)90011-X . free . Lieb . Elliott H. . Elliott H. Lieb . Convex trace functions and the Wigner–Yanase–Dyson conjecture . Advances in Mathematics. 11 . 267–288 . 1973. 3.
  8. 10.1007/BF01646492 . H. Epstein . Remarks on two theorems of E. Lieb . Communications in Mathematical Physics. 31. 317–325 . 1973. 4 . 1973CMaPh..31..317E . 120096681 .
  9. Exercises 2.9 and 2.10
  10. 10.1063/1.1705306 . R. M. Wilcox . Exponential Operators and Parameter Differentiation in Quantum Physics . Journal of Mathematical Physics . 8 . 962–982 . 1967. 4. 1967JMP.....8..962W .
  11. Theorem 5.4
  12. Adrian S. . Lewis . Hristo S. . Sendov . Twice differentiable spectral functions . SIAM Journal on Matrix Analysis and Applications . 23 . 2 . 368–386 . 2001 . 10.1137/S089547980036838X . See Theorem 3.3.
  13. Charles-Alban . Deledalle . Loïc . Denis . Florence . Tupin . Speckle reduction in matrix-log domain for synthetic aperture radar imaging . Journal of Mathematical Imaging and Vision . 2022 . 64 . 3 . 298–320 . 10.1007/s10851-022-01067-1 . free. See Propositions 1 and 2.