In the field of mathematics, norms are defined for elements within a vector space. Specifically, when the vector space comprises matrices, such norms are referred to as matrix norms. Matrix norms differ from vector norms in that they must also interact with matrix multiplication.
K
Km
m
n
K
Km
Norms are often expressed with double vertical bars (like so:
\|A\|
\| ⋅ \|:Km\to\R
For all scalars
\alpha\inK
A,B\inKm
\|A\|\ge0
\|A\|=0\iffA=0m,n
\left\|\alphaA\right\|=\left|\alpha\right|\left\|A\right\|
\|A+B\|\le\|A\|+\|B\|
The only feature distinguishing matrices from rearranged vectors is multiplication. Matrix norms are particularly useful if they are also sub-multiplicative:[3]
\left\|AB\right\|\le\left\|A\right\|\left\|B\right\|
Every norm on can be rescaled to be sub-multiplicative; in some books, the terminology matrix norm is reserved for sub-multiplicative norms.[5]
\| ⋅ \|\alpha
Kn
\| ⋅ \|\beta
Km
m x n
Kn
Km
Km
m x n
\sup
A
\| ⋅ \|\alpha
\| ⋅ \|\beta
\| ⋅ \|\alpha,\beta
If the p-norm for vectors (
1\leqp\leqinfty
Kn
Km,
\|A\|p.
Geometrically speaking, one can imagine a p-norm unit ball
Vp,=\{x\inKn:\|x\|p\le1\}
Kn
A
AVp,\subsetKm
\|A\|p
Vp,
Km
\|A\|p
AVp,
When
p=1,infty
When
p=2
\ell2
A
A
A*A,
A*
A
\sigmamax(A)
A.
There are further properties:
A
\|A\|rm{F}
A
\|A\|2=\sqrt{\rho(A*A)}\leq\sqrt{\|A*A\|infty}\leq\sqrt{\|A\|1\|A\|infty}
We can generalize the above definition. Suppose we have vector norms
\| ⋅ \|\alpha
\| ⋅ \|\beta
Kn
Km
\|A\|p
\|A\|p,
In the special cases of
\alpha=2
\beta=infty
Ai:
A
In the special cases of
\alpha=1
\beta=2
A:j
A
Hence,
\|A\|2,infty
\|A\|1,
Any operator norm is consistent with the vector norms that induce it, giving
Suppose
\| ⋅ \|\alpha,\beta
\| ⋅ \|\beta,\gamma
\| ⋅ \|\alpha,\gamma
(\| ⋅ \|\alpha,\| ⋅ \|\beta)
(\| ⋅ \|\beta,\| ⋅ \|\gamma)
(\| ⋅ \|\alpha,\| ⋅ \|\gamma)
\|AB\|\alpha,\gamma\leq\|A\|\beta,\|B\|\alpha,;
Suppose
\| ⋅ \|\alpha,
Kn
\| ⋅ \|\alpha
\| ⋅ \|\alpha
Moreover, any such norm satisfies the inequalityfor all positive integers r, where is the spectral radius of . For symmetric or hermitian, we have equality in for the 2-norm, since in this case the 2-norm is precisely the spectral radius of . For an arbitrary matrix, we may not have equality for any norm; a counterexample would bewhich has vanishing spectral radius. In any case, for any matrix norm, we have the spectral radius formula:
A matrix norm
\| ⋅ \|
Km
\| ⋅ \|\alpha
Kn
\| ⋅ \|\beta
Km
A\inKm
x\inKn
\alpha=\beta
\| ⋅ \|
\| ⋅ \|\alpha
All induced norms are consistent by definition. Also, any sub-multiplicative matrix norm on
Kn
Kn
\left\|v\right\|:=\left\|\left(v,v,...,v\right)\right\|
These norms treat an
m x n
m ⋅ n
\|A\|p,p=\|vec(A)\|p=\left(
m | |
\sum | |
i=1 |
n | |
\sum | |
j=1 |
|aij|p\right)1/p
This is a different norm from the induced p-norm (see above) and the Schatten p-norm (see below), but the notation is the same.
The special case p = 2 is the Frobenius norm, and p = ∞ yields the maximum norm.
Let
(a1,\ldots,an)
A
A
L2,1
\|A\|2,1=
n | |
\sum | |
j=1 |
\|aj\|2=
n | |
\sum | |
j=1 |
\left(
m | |
\sum | |
i=1 |
|aij|2
| ||||
\right) |
The
L2,1
For, the
L2,1
Lp,q
\|A\|p,q=
n | |
\left(\sum | |
j=1 |
\left(
m | |
\sum | |
i=1 |
|aij|p
| ||||
\right) |
| ||||
\right) |
.
See main article: Hilbert–Schmidt operator.
See also: Frobenius inner product.
When for the
Lp,q
\|A\|F=
n | |
\sqrt{\sum | |
j |
|aij|2}=\sqrt{\operatorname{trace}\left(A*A\right)}=
min\{m,n\ | |
\sqrt{\sum | |
i=1 |
where the trace is the sum of diagonal entries, and
\sigmai(A)
A
trace(A*A)
A
The Frobenius norm is an extension of the Euclidean norm to
Kn
The Frobenius norm is sub-multiplicative and is very useful for numerical linear algebra. The sub-multiplicativity of Frobenius norm can be proved using Cauchy–Schwarz inequality.
Frobenius norm is often easier to compute than induced norms, and has the useful property of being invariant under rotations (and unitary operations in general). That is,
\|A\|F=\|AU\|F=\|UA\|F
U
\operatorname{trace}(XYZ)=\operatorname{trace}(YZX)=\operatorname{trace}(ZXY)
2 | |
\|AU\| | |
F |
=\operatorname{trace}\left((AU)*AU\right) =\operatorname{trace}\left(U*A*AU\right) =\operatorname{trace}\left(UU*A*A\right) =\operatorname{trace}\left(A*A\right) =
2, | |
\|A\| | |
F |
and analogously:
2 | |
\|UA\| | |
F |
=\operatorname{trace}\left((UA)*UA\right) =\operatorname{trace}\left(A*U*UA\right) =\operatorname{trace}\left(A*A\right) =
2, | |
\|A\| | |
F |
where we have used the unitary nature of
U
U*U=UU*=I
It also satisfies
\|A*A\|F=
*\| | |
\|AA | |
F |
\leq
2 | |
\|A\| | |
F |
\|A+
2 | |
B\| | |
F |
=
2 | |
\|A\| | |
F |
+
2 | |
\|B\| | |
F |
+2Re\left(\langleA,B\rangleF\right),
where
\langleA,B\rangleF
The max norm is the elementwise norm in the limit as goes to infinity:
\|A\|max=maxi,|aij|.
This norm is not sub-multiplicative; but modifying the right-hand side to
\sqrt{mn}maxi,\vertai\vert
Note that in some literature (such as Communication complexity), an alternative definition of max-norm, also called the
\gamma2
\gamma2(A)=
min | |
U,V:A=UVT |
\|U\|2,infty\|V\|2,infty=
min | |
U,V:A=UVT |
maxi,j\|Ui,:\|2\|Vj,:\|2
The Schatten p-norms arise when applying the p-norm to the vector of singular values of a matrix. If the singular values of the
m x n
A
\|A\|p=\left(
min\{m,n\ | |
\sum | |
i=1 |
These norms again share the notation with the induced and entry-wise p-norms, but they are different.
All Schatten norms are sub-multiplicative. They are also unitarily invariant, which means that
\|A\|=\|UAV\|
A
U
V
The most familiar cases are p = 1, 2, ∞. The case p = 2 yields the Frobenius norm, introduced before. The case p = ∞ yields the spectral norm, which is the operator norm induced by the vector 2-norm (see above). Finally, p = 1 yields the nuclear norm (also known as the trace norm, or the Ky Fan 'n'-norm[8]), defined as:
\|A\|*=\operatorname{trace}\left(\sqrt{A*A}\right)=
min\{m,n\ | |
\sum | |
i=1 |
where
\sqrt{A*A}
B
BB=A*A
A*A
\|A\|*
rank(A)
Combining von Neumann's trace inequality with Hölder's inequality for Euclidean space yields a version of Hölder's inequality for Schatten norms for
1/p+1/q=1
\left|\operatorname{trace}(A'B)\right|\le\|A\|p\|B\|q,
In particular, this implies the Schatten norm inequality
2 | |
\|A\| | |
F |
\le\|A\|p\|A\|q.
A matrix norm
\| ⋅ \|
A\preccurlyeqB ⇒ \|A\|\leq\|B\|.
The Frobenius norm and spectral norm are examples of monotone norms.[9]
Another source of inspiration for matrix norms arises from considering a matrix as the adjacency matrix of a weighted, directed graph.[10] The so-called "cut norm" measures how close the associated graph is to being bipartite: where .[11] [12] Equivalent definitions (up to a constant factor) impose the conditions ; ; or .
The cut-norm is equivalent to the induced operator norm, which is itself equivalent to another norm, called the Grothendieck norm.
To define the Grothendieck norm, first note that a linear operator is just a scalar, and thus extends to a linear operator on any . Moreover, given any choice of basis for and, any linear operator extends to a linear operator, by letting each matrix element on elements of via scalar multiplication. The Grothendieck norm is the norm of that extended operator; in symbols:
The Grothendieck norm depends on choice of basis (usually taken to be the standard basis) and .
For any two matrix norms
\| ⋅ \|\alpha
\| ⋅ \|\beta
r\|A\|\alpha\leq\|A\|\beta\leqs\|A\|\alpha
for some positive numbers r and s, for all matrices
A\inKm
Km
Km
Km
m x n
Moreover, for every matrix norm
\| ⋅ \|
\Rn x
k
\ell\| ⋅ \|
\ell\gek
k=\sup\{\VertAB\Vert:\VertA\Vert\leq1,\VertB\Vert\leq1\}
A sub-multiplicative matrix norm
\| ⋅ \|\alpha
\| ⋅ \|\beta
\| ⋅ \|\beta<\| ⋅ \|\alpha
Let
\|A\|p
For matrix
A\in\Rm x
r
\|A\|2\le\|A\|F\le\sqrt{r}\|A\|2
\|A\|F\le\|A\|*\le\sqrt{r}\|A\|F
\|A\|max\le\|A\|2\le\sqrt{mn}\|A\|max
1 | |
\sqrt{n |
1 | |
\sqrt{m |