In statistical decision theory, a randomised decision rule or mixed decision rule is a decision rule that associates probabilities with deterministic decision rules. In finite decision problems, randomised decision rules define a risk set which is the convex hull of the risk points of the nonrandomised decision rules.
As nonrandomised alternatives always exist to randomised Bayes rules, randomisation is not needed in Bayesian statistics, although frequentist statistical theory sometimes requires the use of randomised rules to satisfy optimality conditions such as minimax, most notably when deriving confidence intervals and hypothesis tests about discrete probability distributions.
A statistical test making use of a randomized decision rule is called a randomized test.
Let
lD=\{d1,d2...,dh\}
p1,p2,...,ph
d*
h | |
\sum | |
i=1 |
pidi
R(\theta,d*)
h | |
\sum | |
i=1 |
piR(\theta,di)
d1,...,dh\inlD
p1,...ph
Alternatively, a randomised decision rule may assign probabilities directly on elements of the actions space
lA
d*(x,A)
a\inlA
\intAd*(x,A)L(\theta,A)dA
The introduction of randomised decision rules thus creates a larger decision space from which the statistician may choose his decision. As non-randomised decision rules are a special case of randomised decision rules where one decision or action has probability 1, the original decision space
lD
lD*
As with nonrandomised decision rules, randomised decision rules may satisfy favourable properties such as admissibility, minimaxity and Bayes. This shall be illustrated in the case of a finite decision problem, i.e. a problem where the parameter space is a finite set of, say,
k
lS
(R(\theta1,d*),...R(\thetak,d*)),d*\inlD*
(R(\theta1,d),...R(\thetak,d)),d\inlD
In the case where the parameter space has only two elements
\theta1
\theta2
R2
R1
R2
\theta1
\theta2
An admissible decision rule is one that is not dominated by any other decision rule, i.e. there is no decision rule that has equal risk as or lower risk than it for all parameters and strictly lower risk than it for some parameter. In a finite decision problem, the risk point of an admissible decision rule has either lower x-coordinates or y-coordinates than all other risk points or, more formally, it is the set of rules with risk points of the form
(a,b)
\{(R1,R2):R1\leqa,R2\leqb\}\caplS=(a,b)
A minimax Bayes rule is one that minimises the supremum risk
\sup\thetaR(\theta,d*)
lD*
In a finite decision problem with two possible parameters, the minimax rule can be found by considering the family of squares
Q(c)=\{(R1,R2):0\leqR1\leqc,0\leqR2\leqc\}
c
lS
If the risk set intersects the line
R1=R2
R2>R1
R1>R2
r(\pi,d*)
\pi1R1+(1-\pi1)R2=c
\pi1
\pi2
\theta1
\theta2
c
c
As different priors result in different slopes, the set of all rules that are Bayes with respect to some prior are the same as the set of admissible rules.
Note that no situation is possible where a nonrandomised Bayes rule does not exist but a randomised Bayes rule does. The existence of a randomised Bayes rule implies the existence of a nonrandomised Bayes rule. This is also true in the general case, even with infinite parameter space, infinite Bayes risk, and regardless of whether the infimum Bayes risk can be attained.[3] [12] This supports the intuitive notion that the statistician need not utilise randomisation to arrive at statistical decisions.[4]
As randomised Bayes rules always have nonrandomised alternatives, they are unnecessary in Bayesian statistics. However, in frequentist statistics, randomised rules are theoretically necessary under certain situations,[13] and were thought to be useful in practice when they were first invented: Egon Pearson forecast that they 'will not meet with strong objection'. However, few statisticians actually implement them nowadays.[14]
Randomized tests should not be confused with permutation tests.
In the usual formulation of the likelihood ratio test, the null hypothesis is rejected whenever the likelihood ratio
Λ
K
Λ
Λ=K
A solution is to define a test function
\phi(x)
\phi(x)=\left\{\begin{array}{l}1&ifΛ>K\ p(x)&ifΛ=K\ 0&ifΛ<K\end{array}\right.
This can be interpreted as flipping a biased coin with a probability
p(x)
Λ=k
A generalised form of the Neyman–Pearson lemma states that this test has maximum power among all tests at the same significance level
\alpha
\alpha
As an example, consider the case where the underlying distribution is Bernoulli with probability
p
p\leqλ
p>λ
k
P(\hat{p}>k|H0)=\alpha
\hat{p}>k
\hat{p}
\hatp=k
\phi(x)=\left\{\begin{array}{l}1&if\hat{p}>k\ \gamma&if\hat{p}=k\ 0&if\hat{p}<k\end{array}\right.
where
\gamma
P(\hat{p}>k|H0)+\gammaP(\hat{p}=k|H0)=\alpha
An analogous problem arises in the construction of confidence intervals. For instance, the Clopper-Pearson interval is always conservative because of the discrete nature of the binomial distribution. An alternative is to find the upper and lower confidence limits
U
L
\left\{\begin{array}{l}Pr(\hat{p}<k|p=U)+\gammaP(\hat{p}=k|p=U)&=\alpha/2\\ Pr(\hat{p}>k|p=L)+\gammaP(\hat{p}=k|p=L)&=\alpha/2\end{array}\right.
where
\gamma