Control variates explained
The control variates method is a variance reduction technique used in Monte Carlo methods. It exploits information about the errors in estimates of known quantities to reduce the error of an estimate of an unknown quantity.[1] [2] [3]
Underlying principle
Let the unknown parameter of interest be
, and assume we have a
statistic
such that the
expected value of
m is μ:
, i.e.
m is an
unbiased estimator for μ. Suppose we calculate another statistic
such that
is a known value. Then
m\star=m+c\left(t-\tau\right)
is also an unbiased estimator for
for any choice of the coefficient
. The
variance of the resulting estimator
is
rm{Var}\left(m\star\right)=rm{Var}\left(m\right)+c2rm{Var}\left(t\right)+2crm{Cov}\left(m,t\right).
By differentiating the above expression with respect to
, it can be shown that choosing the optimal coefficient
c\star=-
| rm{Cov |
\left(m,t\right)}{rm{Var}\left(t\right)} |
minimizes the variance of
. (Note that this coefficient is the same as the coefficient obtained from a
linear regression.) With this choice,
\begin{align}
rm{Var}\left(m\star\right)&=rm{Var}\left(m\right)-
| \left[rm{Cov |
\left(m,t\right)\right] |
2}{rm{Var}\left(t\right)}\\
&=
| 2\right)rm{Var}\left(m\right)
\end{align} |
\left(1-\rho | |
| m,t |
where
\rhom,t=rm{Corr}\left(m,t\right)
is the correlation coefficient of
and
. The greater the value of
, the greater the
variance reduction achieved.
In the case that
,
, and/or
are unknown, they can be estimated across the Monte Carlo replicates. This is equivalent to solving a certain
least squares system; therefore this technique is also known as
regression sampling.
When the expectation of the control variable,
, is not known analytically, it is still possible to increase the precision in estimating
(for a given fixed simulation budget), provided that the two conditions are met: 1) evaluating
is significantly cheaper than computing
; 2) the magnitude of the correlation coefficient
is close to unity.
[3] Example
We would like to estimate
using
Monte Carlo integration. This integral is the expected value of
, where
and
U follows a
uniform distribution [0, 1].Using a sample of size
n denote the points in the sample as
. Then the estimate is given by
Now we introduce
as a control variate with a known expected value
| 1 |
E\left[g\left(U\right)\right]=\int | |
| 0 |
(1+x)dx=\tfrac{3}{2}
and combine the two into a new estimate
I ≈
\sumi
\sumig(ui)-3/2\right).
Using
realizations and an estimated optimal coefficient
we obtain the following results
Estimate | Variance |
Classical estimate | 0.69475 | 0.01947 |
Control variates | 0.69295 | 0.00060 | |
The variance was significantly reduced after using the control variates technique. (The exact result is
.)
See also
Notes
- Lemieux . C.. Control Variates. Wiley StatsRef: Statistics Reference Online. 2017. 1–8. 10.1002/9781118445112.stat07947 . 9781118445112 .
- Glasserman, P. (2004). Monte Carlo Methods in Financial Engineering. New York: Springer. (p. 185)
- Botev. Z.. Ridder. A.. Variance Reduction. Wiley StatsRef: Statistics Reference Online. 2017. 1–6. 10.1002/9781118445112.stat07975. 9781118445112 .
References
- Ross, Sheldon M. (2002) Simulation 3rd edition
- Averill M. Law & W. David Kelton (2000), Simulation Modeling and Analysis, 3rd edition.
- S. P. Meyn (2007) Control Techniques for Complex Networks, Cambridge University Press. . Downloadable draft (Section 11.4: Control variates and shadow functions)