In statistics, and specifically in the study of the Dirichlet distribution, a neutral vector of random variables is one that exhibits a particular type of statistical independence amongst its elements.[1] In particular, when elements of the random vector must add up to certain sum, then an element in the vector is neutral with respect to the others if the distribution of the vector created by expressing the remaining elements as proportions of their total is independent of the element that was omitted.
A single element
Xi
X1,X2,\ldots,Xk
Xi
Formally, consider the vector of random variables
X=(X1,\ldots,Xk)
k | |
\sum | |
i=1 |
Xi=1.
Xi
X1
X
X1
X1
* | |
X | |
1 |
=\left(
X2 | |
1-X1 |
,
X3 | |
1-X1 |
,\ldots,
Xk | |
1-X1 |
\right).
Variable
X2
X2/(1-X1)
X2/(1-X1)
* | |
X | |
1,2 |
=\left(
X3 | |
1-X1-X2 |
,
X4 | |
1-X1-X2 |
,\ldots,
Xk | |
1-X1-X2 |
\right).
Thus
X2
Y=(X2,X3,\ldots,Xk)
In general, variable
Xj
X1,\ldotsXj-1
* | |
X | |
1,\ldots,j |
=\left(
Xj+1 | |
1-X1- … -Xj |
,\ldots,
Xk | |
1-X1- … -Xj |
\right).
A vector for which each element is neutral is completely neutral.
If
X=(X1,\ldots,XK)\sim\operatorname{Dir}(\alpha)
X