Multinomial test is the statistical test of the null hypothesis that the parameters of a multinomial distribution equal specified values; it is used for categorical data.
Beginning with a sample of
~N~
k
~x=(x1,x2,...,xk)~
~
k | |
\sum | |
i=1 |
xi=N~.
Next, defining a vector of parameters
~H0:\boldsymbol{\pi}=(\pi1,\pi2,\ldots,\pik)~,
~
k | |
\sum | |
i=1 |
\pii=1~.
The exact probability of the observed configuration
~x~
~\operatornameP\left(x\right)0=N!
k | |
\prod | |
i=1 |
| |||||||
xi! |
~.
The significance probability for the test is the probability of occurrence of the data set observed, or of a data set less likely than that observed, if the null hypothesis is true. Using an exact test, this is calculated as
~pl{[sig]}=\sum\left(y\right)\le\operatorname{P
where the sum ranges over all outcomes as likely as, or less likely than, that observed. In practice this becomes computationally onerous as
~k~
~N~
One of these approximations is the likelihood ratio. An alternative hypothesis can be defined under which each value
~\pii~
~pi=
xi | |
N |
~.
~x~
~\operatornameP\left(x\right)A=N!
k | |
\prod | |
i=1 |
| ||||||||||
xi! |
~.
The natural logarithm of the likelihood ratio,
~[l{LR}]~,
~-2~,
~-2ln([l{LR}])=-2
k | |
\sum | |
i=1 |
xiln\left(
\pii | |
pi |
\right)~.
(The factor
~-2~
If the null hypothesis is true, then as
~N~
~-2ln([l{LR}])~
~k-1~
~-2ln([l{LR}])~
~N-1~.
~N-2~
~q1=1+
| ||||||||||||||||
6N(k-1) |
~.
In the special case where the null hypothesis is that all the values
\pii
~1/k~
~q1=1+
k+1 | |
6N |
~.
Subsequently, Smith et al. derived a dividing factor which matches the first moment as far as
~N-3~.
~\pii~,
~q2=1+
k+1 | + | |
6N |
k2 | |
6N2 |
~.
The null hypothesis can also be tested by using Pearson's chi-squared test
~\chi2=
k | |
\sum | |
i=1 |
| |||||||||||||
Ei |
~
where
~Ei=N\pii~
~i~
~k-1~
~-2ln([l{LR}])~
~-2ln([l{LR}])~