In physics, especially in multilinear algebra and tensor analysis, covariance and contravariance describe how the quantitative description of certain geometric or physical entities changes with a change of basis. [1] Briefly, a contravariant vector is a list of numbers that transforms oppositely to a change of basis, and a covariant vector is a list of numbers that transforms in the same way. Contravariant vectors are often just called vectors and covariant vectors are called covectors or dual vectors. The terms covariant and contravariant were introduced by James Joseph Sylvester in 1851.[2] [3]
Curvilinear coordinate systems, such as cylindrical or spherical coordinates, are often used in physical and geometric problems. Associated with any coordinate system is a natural choice of coordinate basis for vectors based at each point of the space, and covariance and contravariance are particularly important for understanding how the coordinate description of a vector changes by passing from one coordinate system to another. Tensors are objects in multilinear algebra that can have aspects of both covariance and contravariance.
In physics, a vector typically arises as the outcome of a measurement or series of measurements, and is represented as a list (or tuple) of numbers such as
(v1,v2,v3).
The numbers in the list depend on the choice of coordinate system. For instance, if the vector represents position with respect to an observer (position vector), then the coordinate system may be obtained from a system of rigid rods, or reference axes, along which the components v1, v2, and v3 are measured. For a vector to represent a geometric object, it must be possible to describe how it looks in any other coordinate system. That is to say, the components of the vectors will transform in a certain way in passing from one coordinate system to another.
A simple illustrative case is that of a Euclidean vector. For a vector, once a set of basis vectors has been defined, then the components of that vector will always vary opposite to that of the basis vectors. That vector is therefore defined as a contravariant tensor. Take a standard position vector for example. By changing the scale of the reference axes from meters to centimeters (that is, dividing the scale of the reference axes by 100, so that the basis vectors now are
.01
A vector, which is an example of a contravariant tensor, has components that transform inversely to the transformation of the reference axes, (with example transformations including rotation and dilation). The vector itself does not change under these operations; instead, the components of the vector change in a way that cancels the change in the spatial axes. In other words, if the reference axes were rotated in one direction, the component representation of the vector would rotate in exactly the opposite way. Similarly, if the reference axes were stretched in one direction, the components of the vector, would reduce in an exactly compensating way. Mathematically, if the coordinate system undergoes a transformation described by an
n x n
\begin{bmatrix}
\prime | |
e | |
1 |
\prime | |
e | |
2 |
...
\prime\end{bmatrix}=\begin{bmatrix} | |
e | |
n |
e1 e2 ... en\end{bmatrix}M
vi
\begin{bmatrix}v1{\prime}\ v2{\prime}\ ...\ vn{\prime}\end{bmatrix}=M-1\begin{bmatrix}v1\ v2\ ...\ vn\end{bmatrix}
By contrast, a covector has components that transform like the reference axes. It lives in the dual vector space, and represents a linear map from vectors to scalars. The dot product operator involving vectors is a good example of a covector. To illustrate, assume we have a covector defined as
v ⋅
v
\begin{bmatrix}v ⋅ e1&v ⋅ e2&...&v ⋅ en\end{bmatrix}
\begin{bmatrix}e1 e2 ... en\end{bmatrix}
w
\begin{bmatrix}w1\ w2\ ...\ wn\end{bmatrix}
n x n
\begin{bmatrix}
\prime | |
e | |
1 |
\prime | |
e | |
2 |
...
\prime\end{bmatrix}=\begin{bmatrix} | |
e | |
n |
e1 e2 ... en\end{bmatrix}M
v ⋅
M
\begin{bmatrix}
\prime | |
v ⋅ e | |
1 |
&
\prime | |
v ⋅ e | |
2 |
&...&
\prime | |
v ⋅ e | |
n |
\end{bmatrix}=\begin{bmatrix}v ⋅ e1&v ⋅ e2&...&v ⋅ en\end{bmatrix}M
A third concept related to covariance and contravariance is invariance. A scalar (also called type-0 or rank-0 tensor) is an object that does not vary with the change in basis. An example of a physical observable that is a scalar is the mass of a particle. The single, scalar value of mass is independent to changes in basis vectors and consequently is called invariant. The magnitude of a vector (such as distance) is another example of an invariant, because it remains fixed even if geometrical vector components vary. (For example, for a position vector of length
3
1
.01
3
100
Thus, to summarize:
v=viei
w=wiei.
iw | |
v | |
i |
The general formulation of covariance and contravariance refers to how the components of a coordinate vector transform under a change of basis (passive transformation). Thus let V be a vector space of dimension n over a field of scalars S, and let each of and be a basis of V.[4] Also, let the change of basis from f to f′ be given by
for some invertible n×n matrix A with entries
i | |
a | |
j |
Yj=\sumi
i | |
a | |
jX |
i.
A vector
v
Xi
where v['''f'''] are elements of the field S known as the components of v in the f basis. Denote the column vector of components of v by v['''f''']:
v[f]=\begin{bmatrix}v1[f]\\v2[f]\\\vdots\\vn[f]\end{bmatrix}
so that can be rewritten as a matrix product
v=fv[f].
The vector v may also be expressed in terms of the f′ basis, so that
v=f'v[f'].
However, since the vector v itself is invariant under the choice of basis,
fv[f]=v=f'v[f'].
The invariance of v combined with the relationship between f and f′ implies that
fv[f]=fAv[fA],
giving the transformation rule
v[f']=v[fA]=A-1v[f].
In terms of components,
vi[fA]=\sumj
j[f] | |
\tilde{a} | |
jv |
where the coefficients
i | |
\tilde{a} | |
j |
Because the components of the vector v transform with the inverse of the matrix A, these components are said to transform contravariantly under a change of basis.
The way A relates the two pairs is depicted in the following informal diagram using an arrow. The reversal of the arrow indicates a contravariant change:
\begin{align} f&\longrightarrowf'\\ v[f]&\longleftarrowv[f'] \end{align}
See main article: Covariant transformation. A linear functional α on V is expressed uniquely in terms of its components (elements in S) in the f basis as
\alpha(Xi)=\alphai[f], i=1,2,...,n.
These components are the action of α on the basis vectors Xi of the f basis.
Under the change of basis from f to f′ (via), the components transform so that
Denote the row vector of components of α by α['''f''']:
\alpha[f]=\begin{bmatrix}\alpha1[f],\alpha2[f],...,\alphan[f]\end{bmatrix}
so that can be rewritten as the matrix product
\alpha[fA]=\alpha[f]A.
Because the components of the linear functional α transform with the matrix A, these components are said to transform covariantly under a change of basis.
The way A relates the two pairs is depicted in the following informal diagram using an arrow. A covariant relationship is indicated since the arrows travel in the same direction:
\begin{align} f&\longrightarrowf'\\ \alpha[f]&\longrightarrow\alpha[f'] \end{align}
Had a column vector representation been used instead, the transformation law would be the transpose
\alphaT[fA]=AT\alphaT[f].
The choice of basis f on the vector space V defines uniquely a set of coordinate functions on V, by means of
xi[f](v)=vi[f].
xi[fA]=
n | |
\sum | |
k=1 |
k[f]. | |
\tilde{a} | |
kx |
This formulation of contravariance and covariance is often more natural in applications in which there is a coordinate space (a manifold) on which vectors live as tangent vectors or cotangent vectors. Given a local coordinate system xi on the manifold, the reference axes for the coordinate system are the vector fields
X1=
\partial | |
\partialx1 |
,...,X | ||||
|
.
If yi is a different coordinate system and
Y | ||||
|
,...,Yn=
\partial | |
\partialyn |
,
f'=fJ-1, J=\left(
\partialyi | |
\partialxj |
n. | |
\right) | |
i,j=1 |
\partial | |
\partialyi |
=
| ||||
\sum | ||||
j=1 |
\partial | |
\partialxj |
.
A tangent vector is by definition a vector that is a linear combination of the coordinate partials
\partial/\partialxi
v=
n | |
\sum | |
i=1 |
vi[f]Xi=f v[f].
Such a vector is contravariant with respect to change of frame. Under changes in the coordinate system, one has
v\left[f'\right]=v\left[fJ-1\right]=Jv[f].
Therefore, the components of a tangent vector transform via
vi\left[f'\right]=
n | |
\sum | |
j=1 |
\partialyi | |
\partialxj |
vj[f].
Accordingly, a system of n quantities vi depending on the coordinates that transform in this way on passing from one coordinate system to another is called a contravariant vector.
In a finite-dimensional vector space V over a field K with a symmetric bilinear form (which may be referred to as the metric tensor), there is little distinction between covariant and contravariant vectors, because the bilinear form allows covectors to be identified with vectors. That is, a vector v uniquely determines a covector α via
\alpha(w)=g(v,w)
Given a basis of V, there is a unique reciprocal basis of V determined by requiring that
i,X | |
g(Y | |
j) |
=
i | |
\delta | |
j, |
\begin{align} v&=\sumi
i[f]X | |
v | |
i |
=fv[f]\\ &=\sumi
i | |
v | |
i[f]Y |
=f\sharpv\sharp[f]. \end{align}
v[fA]=A-1v[f], v\sharp[fA]=ATv\sharp[f].
In the Euclidean plane, the dot product allows for vectors to be identified with covectors. If
e1,e2
e1,e2
\begin{align}
1 ⋅ e | |
e | |
1 |
=1,&
1 ⋅ e | |
e | |
2 |
=0\\
2 ⋅ e | |
e | |
1 |
=0,&
2 ⋅ e | |
e | |
2 |
=1. \end{align}
Thus, e1 and e2 are perpendicular to each other, as are e2 and e1, and the lengths of e1 and e2 normalized against e1 and e2, respectively.
For example,[5] suppose that we are given a basis e1, e2 consisting of a pair of vectors making a 45° angle with one another, such that e1 has length 2 and e2 has length 1. Then the dual basis vectors are given as follows:
Applying these rules, we find
e1=
1 | |
2 |
e1-
1 | |
\sqrt{2 |
e2=-
1 | |
\sqrt{2 |
Thus the change of basis matrix in going from the original basis to the reciprocal basis is
R=\begin{bmatrix}
1 | |
2 |
&-
1 | |
\sqrt{2 |
[e1 e2]=[e1 e2]\begin{bmatrix}
1 | |
2 |
&-
1 | |
\sqrt{2 |
For instance, the vector
v=
3 | |
2 |
e1+2e2
v1=
3 | |
2 |
, v2=2.
The covariant components are obtained by equating the two expressions for the vector v:
v=
1 | |
v | |
1e |
+
2 | |
v | |
2e |
=
1e | |
v | |
1 |
+
2e | |
v | |
2 |
\begin{align} \begin{bmatrix}v1\ v2\end{bmatrix}&=R-1\begin{bmatrix}v1\ v2\end{bmatrix}\\ &=\begin{bmatrix}4&\sqrt{2}\ \sqrt{2}&1\end{bmatrix} \begin{bmatrix}v1\ v2\end{bmatrix}\\ &=\begin{bmatrix}6+2\sqrt{2}\ 2+
3 | |
\sqrt{2 |
In the three-dimensional Euclidean space, one can also determine explicitly the dual basis to a given set of basis vectors e1, e2, e3 of E3 that are not necessarily assumed to be orthogonal nor of unit norm. The dual basis vectors are:
e1=
e2 x e3 | |
e1 ⋅ (e2 x e3) |
; e2=
e3 x e1 | |
e2 ⋅ (e3 x e1) |
; e3=
e1 x e2 | |
e3 ⋅ (e1 x e2) |
.
Even when the ei and ei are not orthonormal, they are still mutually reciprocal:
ei ⋅ ej=
i | |
\delta | |
j, |
Then the contravariant components of any vector v can be obtained by the dot product of v with the dual basis vectors:
q1=v ⋅ e1; q2=v ⋅ e2; q3=v ⋅ e3.
Likewise, the covariant components of v can be obtained from the dot product of v with basis vectors, viz.
q1=v ⋅ e1; q2=v ⋅ e2; q3=v ⋅ e3.
Then v can be expressed in two (reciprocal) ways, viz.
v=qiei=q1e1+q2e2+q3e3.
v=qiei=q1e1+q2e2+q3e3
v=(v ⋅ ei)ei=(v ⋅ ei)ei
qi=v ⋅ ei=(qjej) ⋅ ei=(ej ⋅ ei)qj
qi=v ⋅ ei=(qjej) ⋅ ei=(ej ⋅ ei)qj.
If the basis vectors are orthonormal, then they are the same as the dual basis vectors.
More generally, in an n-dimensional Euclidean space V, if a basis is
e1,...,en,
ei=gijej
gij=ei ⋅ ej.
ij | |
e | |
k=g |
ej ⋅ e
ij | |
k=g |
gjk=
i | |
\delta | |
k |
.
The covariant and contravariant components of any vector
v=qiei=qiei
are related as above by
qi=v ⋅ ei=(qjej) ⋅ ei=
jg | |
q | |
ji |
qi=v ⋅ ei=
j) ⋅ | |
(q | |
je |
ei=
ji | |
q | |
jg |
.
In the field of physics, the adjective covariant is often used informally as a synonym for invariant. For example, the Schrödinger equation does not keep its written form under the coordinate transformations of special relativity. Thus, a physicist might say that the Schrödinger equation is not covariant. In contrast, the Klein–Gordon equation and the Dirac equation do keep their written form under these coordinate transformations. Thus, a physicist might say that these equations are covariant.
Despite this usage of "covariant", it is more accurate to say that the Klein–Gordon and Dirac equations are invariant, and that the Schrödinger equation is not invariant. Additionally, to remove ambiguity, the transformation by which the invariance is evaluated should be indicated.
Because the components of vectors are contravariant and those of covectors are covariant, the vectors themselves are often referred to as being contravariant and the covectors as covariant.
The distinction between covariance and contravariance is particularly important for computations with tensors, which often have mixed variance. This means that they have both covariant and contravariant components, or both vector and covector components. The valence of a tensor is the number of variant and covariant terms, and in Einstein notation, covariant components have lower indices, while contravariant components have upper indices. The duality between covariance and contravariance intervenes whenever a vector or tensor quantity is represented by its components, although modern differential geometry uses more sophisticated index-free methods to represent tensors.
In tensor analysis, a covariant vector varies more or less reciprocally to a corresponding contravariant vector. Expressions for lengths, areas and volumes of objects in the vector space can then be given in terms of tensors with covariant and contravariant indices. Under simple expansions and contractions of the coordinates, the reciprocity is exact; under affine transformations the components of a vector intermingle on going between covariant and contravariant expression.
On a manifold, a tensor field will typically have multiple, upper and lower indices, where Einstein notation is widely used. When the manifold is equipped with a metric, covariant and contravariant indices become very closely related to one another. Contravariant indices can be turned into covariant indices by contracting with the metric tensor. The reverse is possible by contracting with the (matrix) inverse of the metric tensor. Note that in general, no such relation exists in spaces not endowed with a metric tensor. Furthermore, from a more abstract standpoint, a tensor is simply "there" and its components of either kind are only calculational artifacts whose values depend on the chosen coordinates.
The explanation in geometric terms is that a general tensor will have contravariant indices as well as covariant indices, because it has parts that live in the tangent bundle as well as the cotangent bundle.
A contravariant vector is one which transforms like
dx\mu | |
d\tau |
x\mu
\tau
\partial\varphi | |
\partialx\mu |
\varphi
In category theory, there are covariant functors and contravariant functors. The assignment of the dual space to a vector space is a standard example of a contravariant functor. Contravariant (resp. covariant) vectors are contravariant (resp. covariant) functors from a
GL(n)
GL(n)
GL(n)
In differential geometry, the components of a vector relative to a basis of the tangent bundle are covariant if they change with the same linear transformation as a change of basis. They are contravariant if they change by the inverse transformation. This is sometimes a source of confusion for two distinct but related reasons. The first is that vectors whose components are covariant (called covectors or 1-forms) actually pull back under smooth functions, meaning that the operation assigning the space of covectors to a smooth manifold is actually a contravariant functor. Likewise, vectors whose components are contravariant push forward under smooth mappings, so the operation assigning the space of (contravariant) vectors to a smooth manifold is a covariant functor. Secondly, in the classical approach to differential geometry, it is not bases of the tangent bundle that are the most primitive object, but rather changes in the coordinate system. Vectors with contravariant components transform in the same way as changes in the coordinates (because these actually change oppositely to the induced change of basis). Likewise, vectors with covariant components transform in the opposite way as changes in the coordinates.
x\mapstofx.