Exponential Tilting (ET), Exponential Twisting, or Exponential Change of Measure (ECM) is a distribution shifting technique used in many parts of mathematics.The different exponential tiltings of a random variable
X
X
Exponential Tilting is used in Monte Carlo Estimation for rare-event simulation, and rejection and importance sampling in particular.In mathematical finance [1] Exponential Tilting is also known as Esscher tilting (or the Esscher transform), and often combined with indirect Edgeworth approximation and is used in such contexts as insurance futures pricing.[2]
The earliest formalization of Exponential Tilting is often attributed to Esscher[3] with its use in importance sampling being attributed to David Siegmund.[4]
Given a random variable
X
P
f
MX(\theta)=E[e\theta]<infty
P\theta
P\theta(X\indx)=
E[e\thetaI[X\indx]] | |
MX(\theta) |
=e\thetaP(X\indx),
where
\kappa(\theta)
\kappa(\theta)=logE[e\theta]=logMX(\theta).
We call
P\theta(X\indx)=f\theta(x)
the
\theta
X
f\theta(x)\proptoe\thetaf(x)
The exponential tilting of a random vector
X
P\theta(X\indx)=
\thetaTx-\kappa(\theta) | |
e |
P(X\indx),
where
\kappa(\theta)=logE[\exp\{\thetaTX\}]
The exponentially tilted measure in many cases has the same parametric form as that of
X
For example, in the case of the normal distribution,
N(\mu,\sigma2)
f\theta(x)
N(\mu+\theta\sigma2,\sigma2)
Original distribution[5] [6] | θ-Tilted distribution | |||||||
---|---|---|---|---|---|---|---|---|
Gamma(\alpha,\beta) | Gamma(\alpha,\beta-\theta) | |||||||
Binomial(n,p) | Binomial\left(n,
\right) | |||||||
Poisson(λ) | Poisson(λe\theta) | |||||||
Exponential(λ) | Exponential(λ-\theta) | |||||||
l{N}(\mu,\sigma2) | l{N}(\mu+\theta\sigma2,\sigma2) | |||||||
l{N}(\mu,\Sigma) | l{N}(\mu+\Sigma\theta,\Sigma) | |||||||
\chi2(\kappa) |
,
\right) |
For some distributions, however, the exponentially tilted distribution does not belong to the same parametric family as
f
f(x)=\alpha/(1+x)\alpha,x>0
f\theta(x)
\theta<0
In statistical mechanics, the energy of a system in equilibrium with a heat bath has the Boltzmann distribution:
P(E\indE)\proptoe-\betadE
\beta
P\theta(E\indE)\proptoe-(\betadE
Similarly, the energy and particle number of a system in equilibrium with a heat and particle bath has the grand canonical distribution:
P((N,E)\in(dN,dE))\proptoe\betadNdE
\mu
In many cases, the tilted distribution belongs to the same parametric family as the original. This is particularly true when the original density belongs to the exponential family of distribution. This simplifies random variable generation during Monte-Carlo simulations. Exponential tilting may still be useful if this is not the case, though normalization must be possible and additional sampling algorithms may be needed.
In addition, there exists a simple relationship between the original and tilted CGF,
\kappa\theta(η)=
ηX | |
log(E | |
\theta[e |
])=\kappa(\theta+η)-\kappa(\theta).
We can see this by observing that
F\theta(x)=
x\exp\{\theta | |
\int\limits | |
infty |
y-\kappa(\theta)\}f(y)dy.
Thus,
\begin{align} \kappa\theta(η)&=log\inteηdF\theta(x)\ &=log\inteηe\thetadF(x)\\ &=logE[e(η+\theta)X-\kappa(\theta)]\\ &=log(e\kappa(η+\theta)-\kappa(\theta))\\ &=\kappa(η+\theta)-\kappa(\theta) \end{align}
Clearly, this relationship allows for easy calculation of the CGF of the tilted distribution and thus the distributions moments. Moreover, it results in a simple form of the likelihood ratio. Specifically,
\ell=
dP | |
dP\theta |
=
f(x) | |
f\theta(x) |
=e-
\kappa(η)=logE[\exp(ηX)]
X
\theta
X
\kappa\theta(η)=\kappa(\theta+η)-\kappa(\theta).
This means that the
i
X
\kappa(i)(\theta)
E\theta[X]=\tfrac{d}{dη}\kappa\theta(η)|η=0=\kappa'(\theta)
The variance of the tilted distribution is
2}{dη | |
Var | |
\theta[X]=\tfrac{d |
2}\kappa | |
\theta(η)| |
η=0=\kappa''(\theta)
\theta1
\theta2
\theta1+\theta2
X
X1,X2,...
\theta
X
X1,X2,...
\theta
DKL(P\parallelP\theta)=E\left[log\tfrac{P}{P\theta}\right]
between the tilted distribution
P\theta
P
X
E\theta[X]=\kappa'(\theta)
DKL(P\theta\parallelP)=E\theta\left[log\tfrac{P\theta}{P}\right]=\theta\kappa'(\theta)-\kappa(\theta)
The exponential tilting of
X
X|X\inA
\theta
P\theta
The saddlepoint approximation method is a density approximation methodology often used for the distribution of sums and averages of independent, identically distributed random variables that employs Edgeworth series, but which generally performs better at extreme values. From the definition of the natural exponential family, it follows that
f\theta(\bar{x})=f(\bar{x})\exp\{n(\theta\bar{x}-\kappa(\theta))\}
Applying the Edgeworth expansion for
f\theta(\bar{x})
f\theta(\bar{x})=\psi(z)(Var[\bar{X}])-1/2\left\{1+
\rho3(\theta)h3(z) | |
6 |
+
\rho4(\theta)h4(z) | |
24 |
...\right\},
where
\psi(z)
z=
\bar{x | |
- |
\kappa\bar{x
\rhon(\theta)=\kappa(n)(\theta)\{\kappa''(\theta)n/2\}
and
hn
When considering values of
\bar{x}
|z| → infty
hn(z)
\bar{x}
\theta
\kappa'(\theta)=\bar{x}.
This value of
\theta
\theta
f(\bar{x}) ≈ \left(
n | |
2\pi\kappa''(\theta) |
\right)1/2\exp\{n(\kappa(\theta)-\theta\bar{x})\}.
Using the tilted distribution
P\theta
f\theta(x)
1 | |
c |
\exp(-\thetax+\kappa(\theta)),
where
c=\sup\limitsx\in
dP | |
dP\theta |
(x).
That is, a uniformly distributed random variable
p\simUnif(0,1)
f\theta(x)
p\leq
1 | |
c |
\exp(-\thetax+\kappa(\theta)).
Applying the exponentially tilted distribution as the importance distribution yields the equation
E(h(X))=E\theta[\ell(X)h(X)]
where
\ell(X)=
dP | |
dP\theta |
is the likelihood function. So, one samples from
f\theta
P(dX)
Var(X)=E[(\ell(X)h(X)2]
Assume independent and identically distributed
\{Xi\}
\kappa(\theta)<infty
P(X1+ … +Xn>c)
h(X)=
n | |
I(\sum | |
i=1 |
Xi>c)
The constant
c
na
a
n | |
P(\sum | |
i=1 |
Xi>na)=
E | |
\thetaa |
\left[\exp\{-\thetaa\sum
n | |
i=1 |
Xi+n\kappa(\thetaa)\}I(\sum
n | |
i=1 |
Xi>na)\right]
where
\thetaa
\theta
\kappa'(\thetaa)=a
Given the tilting of a normal R.V., it is intuitive that the exponential tilting of
Xt
\mu
\sigma2
\mu+\theta\sigma2
\sigma2
P
P | |
\theta* |
Xt=Bt+\mut
f(Xt)=
f | |
\theta* |
(X | ||||||||
|
=f(Bt)\exp\{\muBT-
1 | |
2 |
\mu2T\}
\exp\{\muBT-
1 | |
2 |
\mu2T\}
MT
P | |
\theta* |
dX(t)=\mu(t)dt+\sigma(t)dB(t)
dX\theta(t)=\mu\theta(t)dt+\sigma(t)dB(t)
\mu\theta(t)
\mu(t)+\theta\sigma(t)
dP | |
dP\theta |
=
| ||||
\exp\{-\int\limits | ||||
0 |
dB(t)+
| ||||
\int\limits | ||||
0 |
)dt\}
Tilting can also be useful for simulating a process
X(t)
dX(t)=\mu(X(t))dt+dB(t)
X(t)
t | |
\int\limits | |
0 |
dX(t)+X(0)
Pproposal
=P | |
\theta* |
| |||||
dP |
(dX(s):0\leqs\leqt)=
\prod\limits\tau\geq\exp\{\mu(X(\tau))dX(\tau)-
\mu(X(\tau))2 | |
2 |
\}dt=
t\mu(X(\tau))dX(\tau) | |
\exp\{\int\limits | |
0 |
-
| ||||
\int\limits | ||||
0 |
\}dt
M(t)
E[M(t)]=1
fX(t)(y)=
\theta* | |
f | |
X(t) |
(y)E | |
\theta* |
[M(t)|X(t)=y]
fX(t)(y) | |||||||||
|
1 | |
c |
=
1 | |
c |
E | |
\theta* |
[M(t)|X(t)=y]
Assume i.i.d. X's with light tailed distribution and
E[X]>0
\psi(c)=P(\tau(c)<infty)
\tau(c)=
t | |
inf\{t:\sum\limits | |
i=1 |
Xi>c\}
c
\psi(c)
\psi
P\theta(\tau(c)<infty)=1
\theta>\theta0
\theta0
\kappa'(\theta0)=0
\theta=\theta*
\theta*
\kappa(\theta*)=0
\theta*
\underset{x → infty}{\lim\sup}
VarIA(x) | |
PA(x)2 |
<infty
We can only see the input and output of a black box, without knowing its structure. The algorithm is to use only minimal information on its structure. When we generate random numbers, the output may not be within the same common parametric class, such as normal or exponential distributions. An automated way may be used to perform ECM. Let
X1,X2,...
G
X\geq0
ak{F}n=\sigma(X1,...,Xn,U1,...,Un)
U1,U2
X1,X2
\{ak{F}n\}
ak{G}
G
[0,infty)
kG=
infty | |
\int | |
0 |
e\thetaG(dx)<infty
G\theta
dG\theta | |
dG(x) |
=
\thetax-kG | |
e |
\theta
ak{G}
\tau
ak{F}\tau-
Z
Z
G\theta
G\inak{G}
PG(Z<x)=G\theta(x)
x
G
G\theta