In statistics, Dunnett's test is a multiple comparison procedure[1] developed by Canadian statistician Charles Dunnett[2] to compare each of a number of treatments with a single control.[3] [4] Multiple comparisons to a control are also referred to as many-to-one comparisons.
Dunnett's test was developed in 1955;[5] an updated table of critical values was published in 1964.[6]
See main article: Multiple comparisons problem. The multiple comparisons, multiplicity or multiple testing problem occurs when one considers a set of statistical inferences simultaneously or infers a subset of parameters selected based on the observed values. The major issue in any discussion of multiple-comparison procedures is the question of the probability of Type I errors. Most differences among alternative techniques result from different approaches to the question of how to control these errors. The problem is in part technical; but it is really much more a subjective question of how you want to define the error rate and how large you are willing to let the maximum possible error rate be.[7] Dunnett's test are well known and widely used in multiple comparison procedure for simultaneously comparing, by interval estimation or hypothesis testing, all active treatments with a control when sampling from a distribution where the normality assumption is reasonable. Dunnett's test is designed to hold the family-wise error rate at or below
\alpha
The original work on Multiple Comparisons problem was made by Tukey and Scheffé. Their method was a general one, which considered all kinds of pairwise comparisons.[7] Tukey's and Scheffé's methods allow any number of comparisons among a set of sample means. On the other hand, Dunnett's test only compares one group with the others, addressing a special case of multiple comparisons problem—pairwise comparisons of multiple treatment groups with a single control group. In the general case, where we compare each of the pairs, we make
k(k-1)/2
(k-1)
Dunnett's test is performed by computing a Student's t-statistic for each experimental, or treatment, group where the statistic compares the treatment group to a single control group.[8] [9] Since each comparison has the same control in common, the procedure incorporates the dependencies between these comparisons. In particular, the t-statistics are all derived from the same estimate of the error variance which is obtained by pooling the sums of squares for error across all (treatment and control) groups. The formal test statistic for Dunnett's test is either the largest in absolute value of these t-statistics (if a two-tailed test is required), or the most negative or most positive of the t-statistics (if a one-tailed test is required).
In Dunnett's test we can use a common table of critical values, but more flexible options are nowadays readily available in many statistics packages. The critical values for any given percentage point depend on: whether a one- or- two-tailed test is performed; the number of groups being compared; the overall number of trials.
The analysis considers the case where the results of the experiment are numerical, and the experiment is performed to compare p treatments with a control group. The results can be summarized as a set of
(p+1)
(\bar{X0
(\bar{X1
\bar{X0
s
p+1
\bar{Xi
p+1
\sigma2
\mui
s2
\sigma2
Dunnett's test's calculation is a procedure that is based on calculating confidence statements about the true or the expected values of the
p
\bar{Xi
p
\bar{Xi
P
P
P
First, we will denote the available N observations by
Xij
i=1,...,p
j=1,...,Ni
s2=
| ||||||||||||||||
) |
2}{n}
\bar{Xi
i
Ni
i
p | |
n=\sum | |
i=0 |
Ni-(p+1)
mi-m0,(i=1,...,p)
p
mi-m0
P
We will consider the general case where there are
p
zi=\cfrac{\bar{Xi
Di=\cfrac{\bar{Xi
we will also write:
Di=
zi | |
s |
P
p
mi-m0,(i=1,...,p)
\bar{Xi
and the
p
di'
Prob(t1<d1',...,tp<dp')=P
\bar{Xi
For bounding
mi-m0
\bar{Xi
when
di''
Prob(|t1|<d1',...,|tp|<dp')=P
di''
di'
The following example was adapted from one given by Villars and was presented in Dunnett's original paper.[5] The data represent measurements on the breaking strength of fabric treated by three different chemical processes compared with a standard method of manufacture.[10]
Standard | Process 1 | Process 2 | Process 3 | ||
---|---|---|---|---|---|
1 | 55 | 55 | 55 | 50 | |
2 | 47 | 64 | 49 | 44 | |
3 | 48 | 64 | 52 | 41 | |
Means | 50 | 61 | 52 | 45 | |
Variance | 19 | 27 | 9 | 21 |
Dunnett's Test can be calculated by applying the following steps:
1. Input Data with Means and Variances:
2. Calculate Pooled Variance
s2
552+472+482+552+\ldots+412-3(502+612+522+452) | |
8 |
=
152 | |
8 |
=19
3. Calculate Standard Deviation
s
s=\sqrt{19}=4.36
4. Calculate Standard Error:
s\sqrt{
2 | |
N |
5. Determine Critical Value
t
t
For
p=3
d.f.=8
One-sided:
t=2.42
Two-sided:
t=2.88
6. the quantity which must be added to and/or subtracted from the observed differences between the means to give their confidence limits is denoted as
A
t
One-sided:
A=2.42 x 3.56=8.61
Two-sided:
A=2.88 x 3.56=10.25
7. Compute Confidence Limits:
One-sided Limits:
Process 1:
61-50-8.61=2.39lbs
Process 2:
52-50-8.61=-6.61lbs
Process 3:
45-50-8.61=-13.61lbs
Two-sided Limits:
Process 1:
61-50\pm10.25=[0.75,21.25]lbs
Process 2:
52-50\pm10.25=[-8.25,12.25]lbs
Process 3:
45-50\pm10.25=[-15.25,5.25]lbs
8. Draw Conclusions:
One-sided:
Process 1: Breaking strength exceeds the standard by at least 2.39 lbs.
Process 2: Breaking strength does not exceed the standard (negative value).
Process 3: Breaking strength does not exceed the standard (negative value).
Two-sided:
Process 1: Breaking strength exceeds the standard by between 0.75 lbs and 21.25 lbs.
Process 2: Breaking strength is between -8.25 lbs and 12.25 lbs (may or may not exceed the standard).
Process 3: Breaking strength is between -15.25 lbs and 5.25 lbs (may or may not exceed the standard).
. Statistics II for Dummies. registration . 186 . dunnett's test developed by. . Wiley . Deborah J. Rumsey . 2009-08-19 . 2012-08-22.