In numerical analysis, the quasi-Monte Carlo method is a method for numerical integration and solving some other problems using low-discrepancy sequences (also called quasi-random sequences or sub-random sequences) to achieve variance reduction. This is in contrast to the regular Monte Carlo method or Monte Carlo integration, which are based on sequences of pseudorandom numbers.
Monte Carlo and quasi-Monte Carlo methods are stated in a similar way.The problem is to approximate the integral of a function f as the average of the function evaluated at a set of points x1, ..., xN:
\int | |
[0,1]s |
f(u){\rmd}u ≈
1 | |
N |
N | |
\sum | |
i=1 |
f(xi).
Since we are integrating over the s-dimensional unit cube, each xi is a vector of s elements. The difference between quasi-Monte Carlo and Monte Carlo is the way the xi are chosen. Quasi-Monte Carlo uses a low-discrepancy sequence such as the Halton sequence, the Sobol sequence, or the Faure sequence, whereas Monte Carlo uses a pseudorandom sequence. The advantage of using low-discrepancy sequences is a faster rate of convergence. Quasi-Monte Carlo has a rate of convergence close to O(1/N), whereas the rate for the Monte Carlo method is O(N−0.5).[1]
The Quasi-Monte Carlo method recently became popular in the area of mathematical finance or computational finance. In these areas, high-dimensional numerical integrals, where the integral should be evaluated within a threshold ε, occur frequently. Hence, the Monte Carlo method and the quasi-Monte Carlo method are beneficial in these situations.
The approximation error of the quasi-Monte Carlo method is bounded by a term proportional to the discrepancy of the set x1, ..., xN. Specifically, the Koksma–Hlawka inequality states that the error
\varepsilon=\left|
\int | |
[0,1]s |
f(u){\rmd}u-
1 | |
N |
N | |
\sum | |
i=1 |
f(xi)\right|
|\varepsilon|\leqV(f)DN,
where V(f) is the Hardy–Krause variation of the function f (see Morokoff and Caflisch (1995) for the detailed definitions). DN is the so-called star discrepancy of the set (x1,...,xN) and is defined as
DN=
\sup | |
Q\subset[0,1]s |
\left|
numberofpointsinQ | |
N |
-\operatorname{volume}(Q)\right|,
where Q is a rectangular solid in [0,1]s with sides parallel to the coordinate axes. The inequality
|\varepsilon|\leqV(f)DN
O\left( | (logN)s |
N |
\right)
O\left( | 1 |
\sqrtN |
\right)
N
log(N)s
For one-dimensional integration, quadrature methods such as the trapezoidal rule, Simpson's rule, or Newton–Cotes formulas are known to be efficient if the function is smooth. These approaches can be also used for multidimensional integrations by repeating the one-dimensional integrals over multiple dimensions. However, the number of function evaluations grows exponentially as s, the number of dimensions, increases. Hence, a method that can overcome this curse of dimensionality should be used for multidimensional integrations. The standard Monte Carlo method is frequently used when the quadrature methods are difficult or expensive to implement.[2] Monte Carlo and quasi-Monte Carlo methods are accurate and relatively fast when the dimension is high, up to 300 or higher.[3]
Morokoff and Caflisch studied the performance of Monte Carlo and quasi-Monte Carlo methods for integration. In the paper, Halton, Sobol, and Faure sequences for quasi-Monte Carlo are compared with the standard Monte Carlo method using pseudorandom sequences. They found that the Halton sequence performs best for dimensions up to around 6; the Sobol sequence performs best for higher dimensions; and the Faure sequence, while outperformed by the other two, still performs better than a pseudorandom sequence.
However, Morokoff and Caflisch gave examples where the advantage of the quasi-Monte Carlo is less than expected theoretically. Still, in the examples studied by Morokoff and Caflisch, the quasi-Monte Carlo method did yield a more accurate result than the Monte Carlo method with the same number of points. Morokoff and Caflisch remark that the advantage of the quasi-Monte Carlo method is greater if the integrand is smooth, and the number of dimensions s of the integral is small.
Lemieux mentioned the drawbacks of quasi-Monte Carlo:[4]
O\left( | (logN)s |
N |
\right)
O\left( | 1 |
\sqrt{N |
s
N
N>2s
V(f)=infty
* | |
D | |
N |
V(f)
Since the low discrepancy sequence are not random, but deterministic, quasi-Monte Carlo method can be seen as a deterministic algorithm or derandomized algorithm. In this case, we only have the bound (e.g., ε ≤ V(f) DN) for error, and the error is hard to estimate. In order to recover our ability to analyze and estimate the variance, we can randomize the method (see randomization for the general idea). The resulting method is called the randomized quasi-Monte Carlo method and can be also viewed as a variance reduction technique for the standard Monte Carlo method.[5] Among several methods, the simplest transformation procedure is through random shifting. Let be the point set from the low discrepancy sequence. We sample s-dimensional random vector U and mix it with . In detail, for each xj, create
yj=xj+U\pmod1
and use the sequence
(yj)
(xj)