Total variation should not be confused with Total variation distance of probability measures.
In mathematics, the total variation identifies several slightly different concepts, related to the (local or global) structure of the codomain of a function or a measure. For a real-valued continuous function f, defined on an interval [''a'', ''b''] ⊂ R, its total variation on the interval of definition is a measure of the one-dimensional arclength of the curve with parametric equation x ↦ f(x), for x ∈ [''a'', ''b'']. Functions whose total variation is finite are called functions of bounded variation.
The concept of total variation for functions of one real variable was first introduced by Camille Jordan in the paper .[1] He used the new concept in order to prove a convergence theorem for Fourier series of discontinuous periodic functions whose variation is bounded. The extension of the concept to functions of more than one variable however is not simple for various reasons.
f
[a,b]\subsetR
b(f)=\sup | |
V | |
l{P |
l{P}=\left\{P=\{x0,...,
x | |
nP |
\}\midPisapartitionof[a,b]\right\}
a=x0<x1<...<
x | |
nP |
=b
[2] Let Ω be an open subset of Rn. Given a function f belonging to L1(Ω), the total variation of f in Ω is defined as
V(f,\Omega):=\sup\left\{\int\Omegaf(x)\operatorname{div}\phi(x)dx\colon\phi\in
1(\Omega,R | |
C | |
c |
n), \Vert
\phi\Vert | |
Linfty(\Omega) |
\le1\right\},
where
1(\Omega,R | |
C | |
c |
n)
\Omega
\Vert \Vert | |
Linfty(\Omega) |
\operatorname{div}
\Omega\subseteqRn
\mu
(X,\Sigma)
\overline{W
\underline{W
\overline{W
\underline{W
clearly
\overline{W
The variation (also called absolute variation) of the signed measure
\mu
|\mu|(E)=\overline{W
and its total variation is defined as the value of this measure on the whole space of definition, i.e.
\|\mu\|=|\mu|(X)
uses upper and lower variations to prove the Hahn - Jordan decomposition: according to his version of this theorem, the upper and lower variation are respectively a non-negative and a non-positive measure. Using a more modern notation, define
\mu+( ⋅ )=\overline{W
\mu-( ⋅ )=-\underline{W
Then
\mu+
\mu-
\mu=\mu+-\mu-
|\mu|=\mu++\mu-
The last measure is sometimes called, by abuse of notation, total variation measure.
If the measure
\mu
\mu
The variation of the complex-valued measure
\mu
|\mu|(E)=\sup\pi\sumA\isin\pi|\mu(A)| \forallE\in\Sigma
where the supremum is taken over all partitions
\pi
E
This definition coincides with the above definition
|\mu|=\mu++\mu-
The variation so defined is a positive measure (see) and coincides with the one defined by when
\mu
\mu
|\mu|(E)=\sup\pi\sumA\isin\pi\|\mu(A)\| \forallE\in\Sigma
where the supremum is as above. This definition is slightly more general than the one given by since it requires only to consider finite partitions of the space
X
See main article: Total variation distance of probability measures. The total variation of any probability measure is exactly one, therefore it is not interesting as a means of investigating the properties of such measures. However, when μ and ν are probability measures, the total variation distance of probability measures can be defined as
\|\mu-\nu\|
(\mu-\nu)(X)=0
\|\mu-\nu\|=|\mu-\nu|(X)=2\sup\left\{\left|\mu(A)-\nu(A)\right|:A\in\Sigma\right\}
and its values are non-trivial. The factor
2
\delta(\mu,\nu)=\sumx\left|\mu(x)-\nu(x)\right| .
It may also be normalized to values in
[0,1]
\delta(\mu,\nu)=
1 | |
2 |
\sumx\left|\mu(x)-\nu(x)\right|
The total variation of a
C1(\overline{\Omega})
f
f
[a,b]\subsetR
f'
b(f) | |
V | |
a |
=\int
b | |
a |
|f'(x)|dx
If
f
b(f) | |
V | |
a |
=|f(a)-f(b)|
For any differentiable function
f
[a,b]
[a,a1],[a1,a2],...,[aN,b]
a<a1<a2< … <aN<b
f
f
[a,b]
b(f) | |
\begin{align} V | |
a |
&=
a1 | |
V | |
a |
(f)+
a2 | |
V | |
a1 |
(f)+ …
b(f)\\[0.3em] &=|f(a)-f(a | |
+V | |
1)|+|f(a |
1)-f(a2)|+ … +|f(aN)-f(b)| \end{align}
Given a
C1(\overline{\Omega})
f
\Omega\subseteqRn
\partial\Omega
C1
f
V(f,\Omega)=\int\Omega\left|\nablaf(x)\right|dx
The first step in the proof is to first prove an equality which follows from the Gauss–Ostrogradsky theorem.
Under the conditions of the theorem, the following equality holds:
\int\Omegaf\operatorname{div}\varphi=-\int\Omega\nablaf ⋅ \varphi
=From the Gauss–Ostrogradsky theorem:
\int\Omega\operatorname{div}R=\int\partial\OmegaR ⋅ n
R:=f\varphi
\int\Omega\operatorname{div}\left(f\varphi\right)= \int\partial\Omega\left(f\varphi\right) ⋅ n
\varphi
\Omega
\int\Omega\operatorname{div}\left(f\varphi\right)=0
\int\Omega
\partial | |
xi |
\left(f\varphii\right)=0
\int\Omega\varphii\partial
xi |
f+
f\partial | |
xi |
\varphii=0
\int\Omega
f\partial | |
xi |
\varphii=-\int\Omega\varphii\partial
xi |
f
\int\Omegaf\operatorname{div}\varphi=-\int\Omega\varphi ⋅ \nablaf
Under the conditions of the theorem, from the lemma we have:
\int\Omegaf\operatorname{div}\varphi =-\int\Omega\varphi ⋅ \nablaf \leq\left|\int\Omega\varphi ⋅ \nablaf\right| \leq\int\Omega\left|\varphi\right| ⋅ \left|\nablaf\right| \leq\int\Omega\left|\nablaf\right|
\varphi
On the other hand, we consider
\thetaN:=-I\left[-N,N\right]I\{\nabla
* | |
\theta | |
N |
\varepsilon
\thetaN
1 | |
C | |
c |
1 | |
C | |
c |
L1
\begin{align} &\limN\toinfty\int\Omega
* | |
f\operatorname{div}\theta | |
N |
\\[4pt] &=\limN\toinfty\int\{\nabla
It can be seen from the proof that the supremum is attained when
\varphi\to
-\nablaf | |
\left|\nablaf\right| |
.
f
The total variation is a norm defined on the space of measures of bounded variation. The space of measures on a σ-algebra of sets is a Banach space, called the ca space, relative to this norm. It is contained in the larger Banach space, called the ba space, consisting of finitely additive (as opposed to countably additive) measures, also with the same norm. The distance function associated to the norm gives rise to the total variation distance between two measures μ and ν.
For finite measures on R, the link between the total variation of a measure μ and the total variation of a function, as described above, goes as follows. Given μ, define a function
\varphi\colonR\toR
\varphi(t)=\mu((-infty,t])~.
\varphi
\|\mu\|TV=\mu+(X)+\mu-(X)~,
(X,\Sigma)
Total variation can be seen as a non-negative real-valued functional defined on the space of real-valued functions (for the case of functions of one variable) or on the space of integrable functions (for the case of functions of several variables). As a functional, total variation finds applications in several branches of mathematics and engineering, like optimal control, numerical analysis, and calculus of variations, where the solution to a certain problem has to minimize its value. As an example, use of the total variation functional is common in the following two kind of problems
One variable
One and more variables
Measure theory