In statistics, a pivotal quantity or pivot is a function of observations and unobservable parameters such that the function's probability distribution does not depend on the unknown parameters (including nuisance parameters).[1] A pivot need not be a statistic — the function and its can depend on the parameters of the model, but its must not. If it is a statistic, then it is known as an ancillary statistic.
More formally,[2] let
X=(X1,X2,\ldots,Xn)
\theta
g(X,\theta)
\theta
g
Pivotal quantities are commonly used for normalization to allow data from different data sets to be compared. It is relatively easy to construct pivots for location and scale parameters: for the former we form differences so that location cancels, for the latter ratios so that scale cancels.
Pivotal quantities are fundamental to the construction of test statistics, as they allow the statistic to not depend on parameters – for example, Student's t-statistic is for a normal distribution with unknown variance (and mean). They also provide one method of constructing confidence intervals, and the use of pivotal quantities improves performance of the bootstrap. In the form of ancillary statistics, they can be used to construct frequentist prediction intervals (predictive confidence intervals).
One of the simplest pivotal quantities is the z-score. Given a normal distribution with mean
\mu
\sigma2
z=
x-\mu | |
\sigma |
,
N(0,1)
N(\mu,\sigma2/n)
z=
\overline{X | |
- |
\mu}{\sigma/\sqrt{n}}
N(0,1).
Given
n
X=(X1,X2,\ldots,Xn)
\mu
\sigma2
g(x,X)=
x-\overline{X | |
\overline{X}=
1 | |
n |
n{X | |
\sum | |
i} |
s2=
1 | |
n-1 |
n{(X | |
\sum | |
i |
-\overline{X})2}
\mu
\sigma2
g(x,X)
x
X
Using
x=\mu
g(\mu,X)
\nu=n-1
\mu
g
g(\mu,X)
\mu
\sigma
X1,\ldots,Xn
This can be used to compute a prediction interval for the next observation
Xn+1;
In more complicated cases, it is impossible to construct exact pivots. However, having approximate pivots improves convergence to asymptotic normality.
Suppose a sample of size
n
(Xi,Yi)'
\rho
An estimator of
\rho
r=
| ||||
\sum |
n | |
i=1 |
(Xi-\overline{X})(Yi-\overline{Y})}{sXsY}
2, | |
s | |
X |
2 | |
s | |
Y |
X
Y
r
\sqrt{n} | r-\rho |
1-\rho2 |
⇒ N(0,1)
However, a variance-stabilizing transformation
z=\rm{tanh}-1r=
12 | |
ln |
1+r | |
1-r |
z
\sqrt{n}(z-\zeta) ⇒ N(0,1)
\zeta={\rmtanh}-1\rho
n
z
r
\operatorname{Var}(z) ≈
1{n-3} | |
See main article: Robust statistics. From the point of view of robust statistics, pivotal quantities are robust to changes in the parameters — indeed, independent of the parameters — but not in general robust to changes in the model, such as violations of the assumption of normality.This is fundamental to the robust critique of non-robust statistics, often derived from pivotal quantities: such statistics may be robust within the family, but are not robust outside it.