A monotonic likelihood ratio in distributions
f(x)
g(x)
The ratio of the density functions above is monotone in the parameter
x ,
f(x) | |
g(x) |
In statistics, the monotone likelihood ratio property is a property of the ratio of two probability density functions (PDFs). Formally, distributions
f(x)
g(x)
foreveryx2>x1,
f(x2) | |
g(x2) |
\geq
f(x1) | |
g(x1) |
that is, if the ratio is nondecreasing in the argument
x
If the functions are first-differentiable, the property may sometimes be stated
\partial | |
\partialx |
\left(
f(x) | |
g(x) |
\right)\geq0
For two distributions that satisfy the definition with respect to some argument
x ,
x~.
T(X) ,
T(X)~.
The MLRP is used to represent a data-generating process that enjoys a straightforward relationship between the magnitude of some observed variable and the distribution it draws from. If
f(x)
g(x)
x
f
g~.
Suppose you are working on a project, and you can either work hard or slack off. Call your choice of effort
e
q~.
q
e
1: Choose effort
e\in\{H,L\}
H
L
2: Observe
q
f(q | e)~.
\operatorname{P}l[ e=H | q r]=
f(q | H) | |
f(q | H)+f(q | L) |
3: Suppose
f(q | e)
1 | |
1+f(q | L)/f(q | H) |
which, thanks to the MLRP, is monotonically increasing in
q
f(q | L) | |
f(q | H) |
q
Hence if some employer is doing a "performance review" he can infer his employee's behavior from the merits of his work.
Statistical models often assume that data are generated by a distribution from some family of distributions and seek to determine that distribution. This task is simplified if the family has the monotone likelihood ratio property (MLRP).
A family of density functions
l\{ f\theta(x) | \theta\in\Theta r\}
\theta
\Theta
T(X)
\theta1<\theta2 ,
| |||||||
|
T(X)~.
Then we say the family of distributions "has MLR in
T(X)
Family | T(X) f\theta(X) | |
---|---|---|
Exponential [λ] | \sumxi | |
Binomial [n,p] | \sumxi | |
Poisson [λ] | \sumxi | |
Normal [\mu,\sigma] | if \sigma \sumxi |
If the family of random variables has the MLRP in
T(X) ,
H0 : \theta\le\theta0
H1 : \theta>\theta0~.
Example: Let
e
y
f(y;e)~.
f
e1,e2 ,
e2>e1
f(y;e2) | |
f(y;e1) |
y~.
Monotone likelihoods are used in several areas of statistical theory, including point estimation and hypothesis testing, as well as in probability models.
One-parameter exponential families have monotone likelihood-functions. In particular, the one-dimensional exponential family of probability density functions or probability mass functions with
f\theta(x)=c(\theta) h(x) \expl( \pi\left(\theta\right) T\left(x\right) r)
T(x) ,
\pi(\theta)
Monotone likelihood functions are used to construct uniformly most powerful tests, according to the Karlin–Rubin theorem.[1] Consider a scalar measurement having a probability density function parameterized by a scalar parameter
\theta ,
\ell(x)=
| |||||||
|
~.
\ell(x)
x ,
\theta1\geq\theta0
x
H1
\varphi(x)=\begin{cases} 1&ifx>x0\\ 0&ifx<x0 \end{cases}
where
x0
\operatorname{E}l\{ \varphi(X) | \theta0 r\}=\alpha
is the UMP test of size
\alpha
H0 : \theta\leq\theta0~~
~~H1:\theta>\theta0~.
Note that exactly the same test is also UMP for testing
H0 : \theta=\theta0~~
~~H1:\theta>\theta0~.
Monotone likelihood-functions are used to construct median-unbiased estimators, using methods specified by Johann Pfanzagl and others.[2] [3] One such procedure is an analogue of the Rao–Blackwell procedure for mean-unbiased estimators: The procedure holds for a smaller class of probability distributions than does the Rao–Blackwell procedure for mean-unbiased estimation but for a larger class of loss functions.[3]
If a family of distributions
f\theta(x)
T(X) ,
\theta
T(X)
x ,
\theta
T(X)
But not conversely: neither monotone hazard rates nor stochastic dominance imply the MLRP.
Let distribution family
f\theta
x ,
\theta1>\theta0
x1>x0 :
| |||||||
|
\geq
| |||||||
|
,
or equivalently:
f | |
\theta1 |
(x1) f
\theta0 |
(x0)\geq
f | |
\theta1 |
(x0) f
\theta0 |
(x1)~.
Integrating this expression twice, we obtain:
1. To x1 x0 \begin{align} &
(x1) f
(x0) dx0\\[6pt] \geq{}&
(x0) f
(x1) dx0 \end{align} integrate and rearrange to obtain
\geq
| 2. From x0 x1 \begin{align} &
(x1) f
(x0) dx1\\[6pt] \geq{}&
(x0) f
(x1) dx1 \end{align} integrate and rearrange to obtain
\geq
|
Combine the two inequalities above to get first-order dominance:
F | |
\theta1 |
(x)\leq
F | |
\theta0 |
(x)~\forallx
Use only the second inequality above to get a monotone hazard rate:
| |||||||
|
\leq
| |||||||
|
~\forallx
The MLR is an important condition on the type distribution of agents in mechanism design and economics of information, where Paul Milgrom defined "favorableness" of signals (in terms of stochastic dominance) as a consequence of MLR.[4] Most solutions to mechanism design models assume type distributions that satisfy the MLR to take advantage of solution methods that may be easier to apply and interpret.