Error exponents in hypothesis testing explained

In statistical hypothesis testing, the error exponent of a hypothesis testing procedure is the rate at which the probabilities of Type I and Type II decay exponentially with the size of the sample used in the test. For example, if the probability of error

Perror

of a test decays as

e-n

, where

n

is the sample size, the error exponent is

\beta

.

Formally, the error exponent of a test is defined as the limiting value of the ratio of the negative logarithm of the error probability to the sample size for large sample sizes:

\limn

-lnPerror
n
. Error exponents for different hypothesis tests are computed using Sanov's theorem and other results from large deviations theory.

Error exponents in binary hypothesis testing

Consider a binary hypothesis testing problem in which observations are modeled as independent and identically distributed random variables under each hypothesis. Let

Y1,Y2,\ldots,Yn

denote the observations. Let

f0

denote the probability density function of each observation

Yi

under the null hypothesis

H0

and let

f1

denote the probability density function of each observation

Yi

under the alternate hypothesis

H1

.

In this case there are two possible error events. Error of type 1, also called false positive, occurs when the null hypothesis is true and it is wrongly rejected. Error of type 2, also called false negative, occurs when the alternate hypothesis is true and null hypothesis is not rejected. The probability of type 1 error is denoted

P(error\midH0)

and the probability of type 2 error is denoted

P(error\midH1)

.

Optimal error exponent for Neyman–Pearson testing

In the Neyman–Pearson version of binary hypothesis testing, one is interested in minimizing the probability of type 2 error

P(error\midH1)

subject to the constraint that the probability of type 1 error

P(error\midH0)

is less than or equal to a pre-specified level

\alpha

. In this setting, the optimal testing procedure is a likelihood-ratio test.[1] Furthermore, the optimal test guarantees that the type 2 error probability decays exponentially in the sample size

n

according to

\limn

-lnP(error\midH1)
n

=D(f0\parallelf1)

.[2] The error exponent

D(f0\parallelf1)

is the Kullback–Leibler divergence between the probability distributions of the observations under the two hypotheses. This exponent is also referred to as the Chernoff–Stein lemma exponent.

Optimal error exponent for average error probability in Bayesian hypothesis testing

In the Bayesian version of binary hypothesis testing one is interested in minimizing the average error probability under both hypothesis, assuming a prior probability of occurrence on each hypothesis. Let

\pi0

denote the prior probability of hypothesis

H0

. In this case the average error probability is given by

Pave=\pi0P(error\midH0)+(1-\pi0)P(error\midH1)

. In this setting again a likelihood ratio test is optimal and the optimal error decays as

\limn

-lnPave
n

=C(f0,f1)

where

C(f0,f1)

represents the Chernoff-information between the two distributions defined as

C(f0,f1)=maxλ\left[-ln\int

λ
(f
0(x))
(1-λ)
(f
1(x))

dx\right]

.

Notes and References

  1. Book: Testing Statistical Hypotheses. 3. 978-0-387-98864-1. Lehmann. E. L.. Erich Leo Lehmann. Joseph P.. Romano. 2005. Springer. New York.
  2. Book: Elements of Information Theory. 2. Cover. Thomas M.. Thomas M. Cover. Joy A.. Thomas. 2006. Wiley-Interscience. New York.