In statistics, cumulative distribution function (CDF)-based nonparametric confidence intervals are a general class of confidence intervals around statistical functionals of a distribution. To calculate these confidence intervals, all that is required is anindependently and identically distributed (iid) sample from the distribution and known bounds on the support of the distribution. The latter requirement simply means that all the nonzero probability mass of the distribution must be contained in some known interval
[a,b]
The intuition behind the CDF-based approach is that bounds on the CDF of a distribution can be translated into bounds on statistical functionals of that distribution. Given an upper and lower bound on the CDF, the approach involves finding the CDFs within the bounds that maximize and minimize the statistical functional of interest.
Unlike approaches that make asymptotic assumptions, including bootstrap approaches and those that rely on the central limit theorem, CDF-based bounds are valid for finite sample sizes. And unlike bounds based on inequalities such as Hoeffding's and McDiarmid's inequalities, CDF-based bounds use properties of the entire sample and thus often produce significantly tighter bounds.
When producing bounds on the CDF, we must differentiate between pointwise and simultaneous bands.
A pointwise CDF bound is one which only guarantees their Coverage probability of
1-\alpha
One method of generating them is based on the Binomial distribution. Considering a single point of a CDF of value
F(xi)
p=F(xi)
n
CDF-based confidence intervals require a probabilistic bound on the CDF of the distribution from which the sample were generated. A variety of methods exist for generating confidence intervals for the CDF of a distribution,
F
x1,\ldots,xn\simF
\hat{F}n(t)=
1 | |
n |
n1\{x | |
\sum | |
i\le |
t\},
where
1\{A\}
F
P(\supx|F(x)-F
-2n\varepsilon2 | |
n(x)|>\varepsilon)\le2e |
.
This can be viewed as a confidence envelope that runs parallel to, and is equally above and below, the empirical CDF.
The interval that contains the true CDF,
F(x)
1-\alpha
Fn(x)-\varepsilon\leF(x)\leFn(x)+\varepsilon where\varepsilon=\sqrt{
| ||||
2n |
The equally spaced confidence interval around the empirical CDF allows for different rates of violations across the support of the distribution. In particular, it is more common for a CDF to be outside of the CDF bound estimated using the Dvoretzky–Kiefer–Wolfowitz inequality near themedian of the distribution than near the endpoints of the distribution. In contrast, the order statistics-based bound introduced by Learned-Miller and DeStefano[3] allows for an equal rateof violation across all of the order statistics. This in turn results in a bound that is tighter near the ends of the support of the distribution and looser in the middle of the support. Other types of bounds can be generated by varying the rate of violation for the order statistics. For example, if a tighter bound on the distribution is desired on the upper portion of the support, a higher rate of violation can be allowed at the upper portion of the support at the expense of having a lower rate of violation, and thus a looser bound, for the lower portion of the support.
Assume without loss of generality that the support of the distribution is contained in
[0,1].
F
F
L(x)
U(x)
E(X)=
1(1-F(x))dx, | |
\int | |
0 |
the confidence interval for the mean can be computed as
1(1-U(x))dx, | |
\left[\int | |
0 |
1(1-L(x))dx | |
\int | |
0 |
\right].
Assume without loss of generality that the support of the distribution of interest,
F
[0,1]
F
E[F']
The CDF-based framework for generating confidence intervals is very general and can be applied to a variety of other statistical functionals including