In linear algebra, the trace of a square matrix, denoted,[1] is defined to be the sum of elements on the main diagonal (from the upper left to the lower right) of . The trace is only defined for a square matrix .
In mathematical physics texts, if = 0 then the matrix is said to be traceless. This is a misnomer, but widely used, such as in the Pauli Matrices.
It can be proven that the trace of a matrix is the sum of its eigenvalues (counted with multiplicities). It can also be proven that for any two matrices and of appropriate sizes. This implies that similar matrices have the same trace. As a consequence one can define the trace of a linear operator mapping a finite-dimensional vector space into itself, since all matrices describing such an operator with respect to a basis are similar.
The trace is related to the derivative of the determinant (see Jacobi's formula).
The trace of an square matrix is defined as[1] [2] [3] where denotes the entry on the row and column of . The entries of can be real numbers, complex numbers, or more generally elements of a field . The trace is not defined for non-square matrices.
Let be a matrix, with
Then
The trace is a linear mapping. That is,for all square matrices and, and all scalars .[3]
A matrix and its transpose have the same trace:[3]
This follows immediately from the fact that transposing a square matrix does not affect elements along the main diagonal.
The trace of a square matrix which is the product of two matrices can be rewritten as the sum of entry-wise products of their elements, i.e. as the sum of all elements of their Hadamard product. Phrased directly, if and are two matrices, then:
If one views any real matrix as a vector of length (an operation called vectorization) then the above operation on and coincides with the standard dot product. According to the above expression, is a sum of squares and hence is nonnegative, equal to zero if and only if is zero.[4] Furthermore, as noted in the above formula, . These demonstrate the positive-definiteness and symmetry required of an inner product; it is common to call the Frobenius inner product of and . This is a natural inner product on the vector space of all real matrices of fixed dimensions. The norm derived from this inner product is called the Frobenius norm, and it satisfies a submultiplicative property, as can be proven with the Cauchy–Schwarz inequality:if and are real positive semi-definite matrices of the same size. The Frobenius inner product and norm arise frequently in matrix calculus and statistics.
The Frobenius inner product may be extended to a hermitian inner product on the complex vector space of all complex matrices of a fixed size, by replacing by its complex conjugate.
The symmetry of the Frobenius inner product may be phrased more directly as follows: the matrices in the trace of a product can be switched without changing the result. If and are and real or complex matrices, respectively, then[3] [5]
This is notable both for the fact that does not usually equal, and also since the trace of either does not usually equal .[6] The similarity-invariance of the trace, meaning that for any square matrix and any invertible matrix of the same dimensions, is a fundamental consequence. This is proved bySimilarity invariance is the crucial property of the trace in order to discuss traces of linear transformations as below.
Additionally, for real column vectors
a\inRn
b\inRn
More generally, the trace is invariant under circular shifts, that is,
This is known as the cyclic property.
Arbitrary permutations are not allowed: in general,
However, if products of three symmetric matrices are considered, any permutation is allowed, since:where the first equality is because the traces of a matrix and its transpose are equal. Note that this is not true in general for more than three factors.
The trace of the Kronecker product of two matrices is the product of their traces:
The following three properties:characterize the trace up to a scalar multiple in the following sense: If
f
f(xy)=f(yx),
f
\operatorname{tr}
For
n x n
f(I)=n
f
Given any matrix, there is
where are the eigenvalues of counted with multiplicity. This holds true even if is a real matrix and some (or all) of the eigenvalues are complex numbers. This may be regarded as a consequence of the existence of the Jordan canonical form, together with the similarity-invariance of the trace discussed above.
When both and are matrices, the trace of the (ring-theoretic) commutator of and vanishes:, because and is linear. One can state this as "the trace is a map of Lie algebras from operators to scalars", as the commutator of scalars is trivial (it is an Abelian Lie algebra). In particular, using similarity invariance, it follows that the identity matrix is never similar to the commutator of any pair of matrices.
Conversely, any square matrix with zero trace is a linear combination of the commutators of pairs of matrices.[8] Moreover, any square matrix with zero trace is unitarily equivalent to a square matrix with diagonal consisting of all zeros.
The trace of an
n x n
A
tn-1
If is a linear operator represented by a square matrix with real or complex entries and if are the eigenvalues of (listed according to their algebraic multiplicities), then
This follows from the fact that is always similar to its Jordan form, an upper triangular matrix having on the main diagonal. In contrast, the determinant of is the product of its eigenvalues; that is,
Everything in the present section applies as well to any square matrix with coefficients in an algebraically closed field.
If is a square matrix with small entries and denotes the identity matrix, then we have approximately
is more general and describes the differential of the determinant at an arbitrary square matrix, in terms of the trace and the adjugate of the matrix.
From this (or from the connection between the trace and the eigenvalues), one can derive a relation between the trace function, the matrix exponential function, and the determinant:
A related characterization of the trace applies to linear vector fields. Given a matrix, define a vector field on by . The components of this vector field are linear functions (given by the rows of). Its divergence is a constant function, whose value is equal to .
By the divergence theorem, one can interpret this in terms of flows: if represents the velocity of a fluid at location and is a region in, the net flow of the fluid out of is given by, where is the volume of .
The trace is a linear operator, hence it commutes with the derivative:
In general, given some linear map (where is a finite-dimensional vector space), we can define the trace of this map by considering the trace of a matrix representation of, that is, choosing a basis for and describing as a matrix relative to this basis, and taking the trace of this square matrix. The result will not depend on the basis chosen, since different bases will give rise to similar matrices, allowing for the possibility of a basis-independent definition for the trace of a linear map.
Such a definition can be given using the canonical isomorphism between the space of linear maps on and, where is the dual space of . Let be in and let be in . Then the trace of the indecomposable element is defined to be ; the trace of a general element is defined by linearity. The trace of a linear map can then be defined as the trace, in the above sense, of the element of corresponding to f under the above mentioned canonical isomorphism. Using an explicit basis for and the corresponding dual basis for, one can show that this gives the same definition of the trace as given above.
The trace can be estimated unbiasedly by "Hutchinson's trick":[9]
Given any matrixUsually, the random vector is sampled from, and any randomW\in\Rn x
withu\in\Rn
, we haveE[uuT]=I
. (Proof: expand the expectation directly.)E[uTWu]=tr(W)
N(0,I)
\{\pmn-1/2\}n
More sophisticated stochastic estimators of trace have been developed.[10]
If a 2 x 2 real matrix has zero trace, its square is a diagonal matrix.
The trace of a 2 × 2 complex matrix is used to classify Möbius transformations. First, the matrix is normalized to make its determinant equal to one. Then, if the square of the trace is 4, the corresponding transformation is parabolic. If the square is in the interval, it is elliptic. Finally, if the square is greater than 4, the transformation is loxodromic. See classification of Möbius transformations.
The trace is used to define characters of group representations. Two representations of a group are equivalent (up to change of basis on) if for all .
The trace also plays a central role in the distribution of quadratic forms.
The trace is a map of Lie algebras
\operatorname{tr}:ak{gl}n\toK
ak{gl}n
K
ak{sl}n
In fact, there is an internal direct sum decomposition
ak{gl}n=ak{sl}n ⊕ K
Formally, one can compose the trace (the counit map) with the unit map
K\toak{gl}n
ak{gl}n\toak{gl}n
In terms of short exact sequences, one haswhich is analogous to(where
K*=K\setminus\{0\}
1/n
ak{gl}n=ak{sl}n ⊕ K
The bilinear form (where, are square matrices)is called the Killing form, which is used for the classification of Lie algebras.
The trace defines a bilinear form:
The form is symmetric, non-degenerate[11] and associative in the sense that:
For a complex simple Lie algebra (such as), every such bilinear form is proportional to each other; in particular, to the Killing form.
Two matrices and are said to be trace orthogonal if
There is a generalization to a general representation
(\rho,ak{g},V)
ak{g}
\rho
\rho:ak{g} → End(V).
trV
End(V)
The concept of trace of a matrix is generalized to the trace class of compact operators on Hilbert spaces, and the analog of the Frobenius norm is called the Hilbert–Schmidt norm.
(en)n
The partial trace is another generalization of the trace that is operator-valued. The trace of a linear operator which lives on a product space is equal to the partial traces over and :
For more properties and a generalization of the partial trace, see traced monoidal categories.
If is a general associative algebra over a field, then a trace on is often defined to be any map which vanishes on commutators; for all . Such a trace is not uniquely defined; it can always at least be modified by multiplication by a nonzero scalar.
A supertrace is the generalization of a trace to the setting of superalgebras.
The operation of tensor contraction generalizes the trace to arbitrary tensors.
Given a vector space, there is a natural bilinear map given by sending to the scalar . The universal property of the tensor product automatically implies that this bilinear map is induced by a linear functional on .[13]
Similarly, there is a natural bilinear map given by sending to the linear map . The universal property of the tensor product, just as used previously, says that this bilinear map is induced by a linear map . If is finite-dimensional, then this linear map is a linear isomorphism. This fundamental fact is a straightforward consequence of the existence of a (finite) basis of, and can also be phrased as saying that any linear map can be written as the sum of (finitely many) rank-one linear maps. Composing the inverse of the isomorphism with the linear functional obtained above results in a linear functional on . This linear functional is exactly the same as the trace.
Using the definition of trace as the sum of diagonal elements, the matrix formula is straightforward to prove, and was given above. In the present perspective, one is considering linear maps and, and viewing them as sums of rank-one maps, so that there are linear functionals and and nonzero vectors and such that and for any in . Then
(S\circT)(u)=\sumi\varphii\left(\sumj\psij(u)wj\right)vi=\sumi\sumj\psij(u)\varphii(wj)vi
\operatorname{tr}(S\circT)=\sumi\sumj\psij(vi)\varphii(wj)=\sumj\sumi\varphii(wj)\psij(vi).
The above proof can be regarded as being based upon tensor products, given that the fundamental identity of with is equivalent to the expressibility of any linear map as the sum of rank-one linear maps. As such, the proof may be written in the notation of tensor products. Then one may consider the multilinear map given by sending to . Further composition with the trace map then results in, and this is unchanged if one were to have started with instead. One may also consider the bilinear map given by sending to the composition, which is then induced by a linear map . It can be seen that this coincides with the linear map . The established symmetry upon composition with the trace map then establishes the equality of the two traces.
For any finite dimensional vector space, there is a natural linear map ; in the language of linear maps, it assigns to a scalar the linear map . Sometimes this is called coevaluation map, and the trace is called evaluation map. These structures can be axiomatized to define categorical traces in the abstract setting of category theory.
then the product isand the traces are .
eij
f\left(eij\right)=0
i ≠ j
f\left(ejj\right)=f\left(e11\right)
More abstractly, this corresponds to the decompositionas
\operatorname{tr}(AB)=\operatorname{tr}(BA)
\operatorname{tr}([A,B])=0
ak{sl}n,
ak{sl}n