In mathematics, tensor calculus, tensor analysis, or Ricci calculus is an extension of vector calculus to tensor fields (tensors that may vary over a manifold, e.g. in spacetime).
Developed by Gregorio Ricci-Curbastro and his student Tullio Levi-Civita,[1] it was used by Albert Einstein to develop his general theory of relativity. Unlike the infinitesimal calculus, tensor calculus allows presentation of physics equations in a form that is independent of the choice of coordinates on the manifold.
Tensor calculus has many applications in physics, engineering and computer science including elasticity, continuum mechanics, electromagnetism (see mathematical descriptions of the electromagnetic field), general relativity (see mathematics of general relativity), quantum field theory, and machine learning.
Working with a main proponent of the exterior calculus Élie Cartan, the influential geometer Shiing-Shen Chern summarizes the role of tensor calculus:[2]
In our subject of differential geometry, where you talk about manifolds, one difficulty is that the geometry is described by coordinates, but the coordinates do not have meaning. They are allowed to undergo transformation. And in order to handle this kind of situation, an important tool is the so-called tensor analysis, or Ricci calculus, which was new to mathematicians. In mathematics you have a function, you write down the function, you calculate, or you add, or you multiply, or you can differentiate. You have something very concrete. In geometry the geometric situation is described by numbers, but you can change your numbers arbitrarily. So to handle this, you need the Ricci calculus.
Tensor notation makes use of upper and lower indexes on objects that are used to label a variable object as covariant (lower index), contravariant (upper index), or mixed covariant and contravariant (having both upper and lower indexes). In fact in conventional math syntax we make use of covariant indexes when dealing with Cartesian coordinate systems
(x1,x2,x3)
Tensor notation allows upper index on an object that may be confused with normal power operations from conventional math syntax.
Tensors notation allows a vector (
\vec{V}
\vec{Z}i
\vec{Z}i
Vi
Vi
\vec{V}=Vi\vec{Z}i=Vi\vec{Z}i
Every vector has two different representations, one referred to as contravariant component (
Vi
\vec{Z}i
Vi
\vec{Z}i
For example, in physics you start with a vector field, you decompose it with respect to the covariant basis, and that's how you get the contravariant coordinates. For orthonormal cartesian coordinates, the covariant and contravariant basis are identical, since the basis set in this case is just the identity matrix, however, for non-affine coordinate system such as polar or spherical there is a need to distinguish between decomposition by use of contravariant or covariant basis set for generating the components of the coordinate system.
\vec{V}=Vi\vec{Z}i
variable | description | Type | |
---|---|---|---|
\vec{V} | vector | invariant | |
Vi | contravariant components (ordered set of scalars) | variant | |
\vec{Z}i | covariant bases (ordered set of vectors) | variant |
\vec{V}=Vi\vec{Z}i
variable | description | type | |
---|---|---|---|
\vec{V} | vector | invariant | |
Vi | covariant components (ordered set of scalars) | variant | |
\vec{Z}i | contravariant bases (ordered set of covectors) | variant |
The metric tensor represents a matrix with scalar elements (
Zij
Zij
Example of lowering index using metric tensor:
Ti=ZijTj
Example of raising index using metric tensor:
Ti=ZijTj
The metric tensor is defined as:
Zij=\vec{Z}i ⋅ \vec{Z}j
Zij=\vec{Z}i ⋅ \vec{Z}j
This means that if we take every permutation of a basis vector set and dotted them against each other, and then arrange them into a square matrix, we would have a metric tensor. The caveat here is which of the two vectors in the permutation is used for projection against the other vector, that is the distinguishing property of the covariant metric tensor in comparison with the contravariant metric tensor.
Two flavors of metric tensors exist: (1) the contravariant metric tensor (
Zij
Zij
ZikZjk=
j | |
\delta | |
i |
\deltaij
\deltaij
\deltaij=\deltaij=
i | |
\delta | |
j |
In addition a tensor can be readily converted from an unbarred(
x
\bar{x}
by use of Jacobian matrix relationships between the barred and unbarred coordinate system (
\bar{J}=J-1
Contravariant vectors are required to obey the laws:
vi=
| ||||
\bar{v} |
\bar{x}r}
\bar{v}i=
| ||||
v |
xr}
Covariant vectors are required to obey the laws:
vi=
\bar{v} | ||||
|
xr}
\bar{v}i=
v | ||||
|
\bar{x}i}
There are two flavors of Jacobian matrix:
1. The J matrix representing the change from unbarred to barred coordinates. To find J, we take the "barred gradient", i.e. partial derivative with respect to
\bar{x}i
J=\bar{\nabla}f(x(\bar{x}))
2. The
\bar{J}
\bar{J}
xi
\bar{J}=\nabla\bar{f}(\bar{x}(x))
Tensor calculus provides a generalization to the gradient vector formula from standard calculus that works in all coordinate systems:
\nablaF=\nablaiF\vec{Z}i
Where:
\nablaiF=
\partialF | |
\partialZi |
In contrast, for standard calculus, the gradient vector formula is dependent on the coordinate system in use (example: Cartesian gradient vector formula vs. the polar gradient vector formula vs. the spherical gradient vector formula, etc.). In standard calculus, each coordinate system has its own specific formula, unlike tensor calculus that has only one gradient formula that is equivalent for all coordinate systems. This is made possible by an understanding of the metric tensor that tensor calculus makes use of.