In probability theory, calculation of the sum of normally distributed random variables is an instance of the arithmetic of random variables.
This is not to be confused with the sum of normal distributions which forms a mixture distribution.
Let X and Y be independent random variables that are normally distributed (and therefore also jointly so), then their sum is also normally distributed. i.e., if
X\simN(\muX,
2) | |
\sigma | |
X |
Y\simN(\muY,
2) | |
\sigma | |
Y |
Z=X+Y,
then
Z\simN(\muX+\muY,
2 | |
\sigma | |
X |
+
2). | |
\sigma | |
Y |
This means that the sum of two independent normally distributed random variables is normal, with its mean being the sum of the two means, and its variance being the sum of the two variances (i.e., the square of the standard deviation is the sum of the squares of the standard deviations).
In order for this result to hold, the assumption that X and Y are independent cannot be dropped, although it can be weakened to the assumption that X and Y are jointly, rather than separately, normally distributed.[1] (See here for an example.)
The result about the mean holds in all cases, while the result for the variance requires uncorrelatedness, but not independence.
\varphiX+Y(t)=\operatorname{E}\left(eit(X+Y)\right)
of the sum of two independent random variables X and Y is just the product of the two separate characteristic functions:
\varphiX(t)=\operatorname{E}\left(eitX\right), \varphiY(t)=\operatorname{E}\left(eitY\right)
of X and Y.
The characteristic function of the normal distribution with expected value μ and variance σ2 is
\varphi(t)=\exp\left(it\mu-{\sigma2t2\over2}\right).
So
\begin{align} \varphiX+Y(t)=\varphiX(t)\varphiY(t)&=\exp\left(it\muX-
2 | |
{\sigma | |
X |
t2\over2}\right)\exp\left(it\muY-
2 | |
{\sigma | |
Y |
t2\over2}\right)\\[6pt] &=\exp\left(it(\muX+\muY)-
2 | |
{(\sigma | |
X |
+
2) | |
\sigma | |
Y |
t2\over2}\right). \end{align}
This is the characteristic function of the normal distribution with expected value
\muX+\muY
2 | |
\sigma | |
Y |
Finally, recall that no two distinct distributions can both have the same characteristic function, so the distribution of X + Y must be just this normal distribution.
For independent random variables X and Y, the distribution fZ of Z = X + Y equals the convolution of fX and fY:
fZ(z)=
infty | |
\int | |
-infty |
fY(z-x)fX(x)dx
Given that fX and fY are normal densities,
\begin{align} fX(x)=l{N}(x;\muX,
2) | |
\sigma | |
X |
=
1 | |
\sqrt{2\pi |
\sigmaX}
| |||||||
e |
\\[5pt] fY(y)=l{N}(y;\muY,
2) | |
\sigma | |
Y |
=
1 | |
\sqrt{2\pi |
\sigmaY}
| |||||||
e |
\end{align}
Substituting into the convolution:
\begin{align} fZ(z) &=
infty | |
\int | |
-infty |
1 | |
\sqrt{2\pi |
\sigmaY} \exp
2 | |
\left[-{(z-x-\mu | |
Y) |
\over
2}\right] | |
2\sigma | |
Y |
1 | |
\sqrt{2\pi |
\sigmaX} \exp
2 | |
\left[-{(x-\mu | |
X) |
\over
2}\right] | |
2\sigma | |
X |
dx\\[6pt] &=
infty | |
\int | |
-infty |
1 | |
\sqrt{2\pi |
\sqrt{2\pi}\sigmaX\sigmaY} \exp\left[-
| ||||||||||||||||
|
\right] dx\\[6pt] &=
infty | |
\int | |
-infty |
1 | |
\sqrt{2\pi |
\sqrt{2\pi}\sigmaX\sigmaY} \exp\left[ -
| ||||||||||||||||||||||||||||
|
\right] dx\\[6pt] &=
infty | |
\int | |
-infty |
1 | |
\sqrt{2\pi |
\sqrt{2\pi}\sigmaX\sigmaY} \exp\left[ -
| |||||||||||||||||||||||||||||||||||||||||||||
|
\right] dx \\[6pt] \end{align}
Defining
\sigmaZ=
2 | |
\sqrt{\sigma | |
X |
+
2} | |
\sigma | |
Y |
\begin{align} fZ(z) &=
infty | |
\int | |
-infty |
1 | |
\sqrt{2\pi |
\sigmaZ}
1 | |
\sqrt{2\pi |
\sigmaX\sigmaY | |
\sigmaZ |
The expression in the integral is a normal density distribution on x, and so the integral evaluates to 1. The desired result follows:
fZ(z)=
1 | |
\sqrt{2\pi |
\sigmaZ}\exp\left[-{(z-(\muX+\mu
2 | |
Y)) |
\over
2 | |
2\sigma | |
Z |
}\right]
It can be shown that the Fourier transform of a Gaussian,
fX(x)=l{N}(x;\muX,
2) | |
\sigma | |
X |
l{F}\{fX\}=FX(\omega)=\exp\left[-j\omega\muX\right]\exp
2\omega | |
\left[-\tfrac{\sigma | |
X |
2}{2}\right]
By the convolution theorem:
\begin{align} fZ(z)&=(fX*fY)(z)\\[5pt] &=l{F}-1\{l{F}\{fX\} ⋅ l{F}\{fY\}\}\\[5pt] &=l{F}-1\{ \exp\left[-j\omega\muX\right]\exp
2\omega | |
\left[-\tfrac{\sigma | |
X |
2}{2}\right] \exp\left[-j\omega\muY\right]\exp
2\omega | |
\left[-\tfrac{\sigma | |
Y |
2}{2}\right] \}\\[5pt] &=l{F}-1\{ \exp\left[-j\omega(\muX+\muY)\right]\exp
2 + | |
\left[-\tfrac{(\sigma | |
X |
2)\omega | |
\sigma | |
Y |
2}{2}\right] \}\\[5pt] &=l{N}(z;\muX+\muY,
2 | |
\sigma | |
X |
+
2) \end{align} | |
\sigma | |
Y |
First consider the normalized case when X, Y ~ N(0, 1), so that their PDFs are
f(x)=
1 | |
\sqrt{2\pi |
g(y)=
1 | |
\sqrt{2\pi |
z\mapsto\intx+yf(x)g(y)dxdy.
The key observation is that the function
f(x)g(y)=
1 | |
2\pi |
-(x2+y2)/2 | |
e |
is radially symmetric. So we rotate the coordinate plane about the origin, choosing new coordinates
x',y'
x'=c
c=c(z)
f(x)g(y)=f(x')g(y')
\intx'\leqf(x')g(y')dx'dy'.
This is easy to integrate; we find that the CDF for Z is
c(z) | |
\int | |
-infty |
f(x')dx'=\Phi(c(z)).
To determine the value
c(z)
(z/2,z/2)
c=\sqrt{(z/2)2+(z/2)2}=z/\sqrt{2}
\Phi(z/\sqrt{2})
Z=X+Y\simN(0,2).
Now, if a, b are any real constants (not both zero) then the probability that
aX+bY\leqz
ax+by=z
z | |
\sqrt{a2+b2 |
aX+bY\simN(0,a2+b2).
Xi\sim
2), | |
N(0,\sigma | |
i |
i=1,...,n,
X1+ … +Xn\simN(0,
2 | |
\sigma | |
1 |
+ … +
2). | |
\sigma | |
n |
Now we are essentially done, because
X\simN(\mu,\sigma2)\Leftrightarrow
1 | |
\sigma |
(X-\mu)\simN(0,1).
Xi\simN(\mui,
2), | |
\sigma | |
i |
i=1,...,n,
n | |
\sum | |
i=1 |
aiXi\sim
n | |
N\left(\sum | |
i=1 |
ai\mui,
n | |
\sum | |
i=1 |
(ai
2 | |
\sigma | |
i) |
\right).
See also: Markov chain central limit theorem. In the event that the variables X and Y are jointly normally distributed random variables, then X + Y is still normally distributed (see Multivariate normal distribution) and the mean is the sum of the means. However, the variances are not additive due to the correlation. Indeed,
\sigmaX+Y=
2+2\rho\sigma | |
\sqrt{\sigma | |
X |
\sigmaY},
where ρ is the correlation. In particular, whenever ρ < 0, then the variance is less than the sum of the variances of X and Y.
Extensions of this result can be made for more than two random variables, using the covariance matrix.
In this case (with X and Y having zero means), one needs to consider
1 | |
2\pi\sigmax\sigmay\sqrt{1-\rho2 |
As above, one makes the substitution
y → z-x
This integral is more complicated to simplify analytically, but can be done easily using a symbolic mathematics program. The probability distribution fZ(z) is given in this case by
f | ||||
|
\sigma+}\exp\left(-
z2 | ||||||
|
\right)
\sigma+=
2+2\rho\sigma | |
\sqrt{\sigma | |
x |
\sigmay}.
If one considers instead Z = X − Y, then one obtains
f | |||||||||||||
|
\sigmaX-Y
2-2\rho\sigma | |
=\sqrt{\sigma | |
x |
\sigmay}.
The standard deviations of each distribution are obvious by comparison with the standard normal distribution.