Inverse probability weighting is a statistical technique for estimating quantities related to a population other than the one from which the data was collected. Study designs with a disparate sampling population and population of target inference (target population) are common in application. There may be prohibitive factors barring researchers from directly sampling from the target population such as cost, time, or ethical concerns. A solution to this problem is to use an alternate design strategy, e.g. stratified sampling. Weighting, when correctly applied, can potentially improve the efficiency and reduce the bias of unweighted estimators.
One very early weighted estimator is the Horvitz–Thompson estimator of the mean.[1] When the sampling probability is known, from which the sampling population is drawn from the target population, then the inverse of this probability is used to weight the observations. This approach has been generalized to many aspects of statistics under various frameworks. In particular, there are weighted likelihoods, weighted estimating equations, and weighted probability densities from which a majority of statistics are derived. These applications codified the theory of other statistics and estimators such as marginal structural models, the standardized mortality ratio, and the EM algorithm for coarsened or aggregate data.
Inverse probability weighting is also used to account for missing data when subjects with missing data cannot be included in the primary analysis.With an estimate of the sampling probability, or the probability that the factor would be measured in another measurement, inverse probability weighting can be used to inflate the weight for subjects who are under-represented due to a large degree of missing data.
The inverse probability weighting estimator can be used to demonstrate causality when the researcher cannot conduct a controlled experiment but has observed data to model. Because it is assumed that the treatment is not randomly assigned, the goal is to estimate the counterfactual or potential outcome if all subjects in population were assigned either treatment.
Suppose observed data are
\{l(Xi,Ai,Y
n | |
i=1 |
X\inRp
A\in\{0,1\}
Y\inR
The goal is to estimate the potential outcome,
Y*l(ar)
a
\mua=EY*(a)
\mua
\{l(Xi,Ai,Y
n | |
i=1 |
IPWE \hat{\mu} a,n =
1 n
n \sum i=1 Yi
1 Ai=a \hat{p n(Ai|Xi)}
\mua=E\left[
1A=aY | |
p(A|X) |
\right]
p(a|x)=
P(A=a,X=x) | |
P(X=x) |
\hat{p}n(a|x)
p(a|x)
IPWE | |
\hat{\mu} | |
a,n |
=
n | |
\sum | |
i=1 |
| |||||||||||
n\hat{p |
n(Ai|Xi)}
Recall the joint probability model
(X,A,Y)\simP
X
A
Y
X
A
x
a
Y(X=x,A=a)=Y(x,a)
\begin{align} Y(x,a)\sim
P(x,a, ⋅ ) | |
\intP(x,a,y)dy |
. \end{align}
We make the following assumptions.
Y=Y*(A)
\{Y*(0),Y*(1)\}\perpA|X
f
g
P(A=a|X=x)=EA[1(A=a)|X=x]>0
a
x
Under the assumptions (A1)-(A3), we will derive the following identities
\begin{align} E\left[Y*(a)\right]=E(X,Y)\left[Y(X,a)\right]=E(X,A,Y)\left[
Y1(A=a) | |
P(A=a|X) |
\right]. … … (*) \end{align}
The first equality follows from the definition and (A1). For the second equality, first use the iterated expectation to write
\begin{align} E(X,Y)\left[Y(X,a)\right]=EX\left[EY\left[Y(X,a)|X\right]\right]. \end{align}
By (A3),
EA[1(A=a)|X]>0
\begin{align} EY\left[Y(X,a)|X\right]&=
EY\left[Y(X,a)|X\right]EA[1(A=a)|X] | |
EA[1(A=a)|X] |
\\ &=
E(A,Y)\left[Y(X,a)1(A=a)|X\right] | |
E[1(A=a)|X] |
… … (A2)\\ &=E(A,Y)\left[
Y(X,a)1(A=a) | |
E[1(A=a)|X] |
|X\right]. … … (denominatorisafunctionofX) \end{align}
Hence integrating out the last expression with respect to
X
Y(X,a)1(A=a)=Y(X,A)1(A=a)=Y1(A=a)
(*)
The Inverse Probability Weighted Estimator (IPWE) is known to be unstable if some estimated propensities are too close to 0 or 1. In such instances, the IPWE is dominated by a small number of subjects with large weights. Recently developed smoothed IPW estimators by employing Rao-Blackwellization, however, reduce the variance of IPWE by up to 7-fold and can also protect the augmented inverse probability weighted estimator from model misspecification.
An alternative estimator is the augmented inverse probability weighted estimator (AIPWE) combines both the properties of the regression based estimator and the inverse probability weighted estimator. It is therefore a 'doubly robust' method in that it only requires either the propensity or outcome model to be correctly specified but not both. This method augments the IPWE to reduce variability and improve estimate efficiency. This model holds the same assumptions as the Inverse Probability Weighted Estimator (IPWE).[2]
AIPWE | ||
\begin{align} \hat{\mu} | &= | |
a,n |
1 | |
n |
| ||||||||||||||
\sum | ||||||||||||||
n |
(Ai|Xi)}-
| |||||||
n(A |
i|Xi)}{\hat{p}n(Ai|Xi)}\hat{Q}n(Xi,a)r)\\ &=
1 | |
n |
| ||||||||
\sum | ||||||||
n |
(Ai|Xi)}Yi+ (1-
| |||||
\hat{p |
n(Ai|Xi)})\hat{Q}n(Xi,a)r)\\ &=
1 | |
n |
nl(\hat{Q} | |
\sum | |
n(X |
i,a)r)+
1 | |
n |
| ||||||||
\sum | ||||||||
n |
(Ai|Xi)}l(Yi-\hat{Q}n(Xi,a)r) \end{align}
With the following notations:
1 | |
Ai=a |
\hat{Q}n(x,a)
Y
X
A
\hat{p}n(Ai|Xi)
AIPWE | |
\hat{\mu} | |
a,n |
The later rearrangement of the formula helps reveal the underlying idea: our estimator is based on the average predicted outcome using the model (i.e.:
1 | |
n |
nl(\hat{Q} | |
\sum | |
n(X |
i,a)r)
1 | |
n |
| ||||||||
\sum | ||||||||
n |
(Ai|Xi)}l(Yi-\hat{Q}n(Xi,a)r)
The "doubly robust" benefit of such an estimator comes from the fact that it's sufficient for one of the two models to be correctly specified, for the estimator to be unbiased (either
\hat{Q}n(Xi,a)
\hat{p}n(Ai|Xi)
The bias of the doubly robust estimators is called a second-order bias, and it depends on the product of the difference
1 | |
\hat{p |
n(Ai|Xi)}-
1 | |
{p |
n(Ai|Xi)}
\hat{Q}n(Xi,a)-Qn(Xi,a)