In mathematics, the moments of a function are certain quantitative measures related to the shape of the function's graph. If the function represents mass density, then the zeroth moment is the total mass, the first moment (normalized by total mass) is the center of mass, and the second moment is the moment of inertia. If the function is a probability distribution, then the first moment is the expected value, the second central moment is the variance, the third standardized moment is the skewness, and the fourth standardized moment is the kurtosis.
For a distribution of mass or probability on a bounded interval, the collection of all the moments (of all orders, from to) uniquely determines the distribution (Hausdorff moment problem). The same is not true on unbounded intervals (Hamburger moment problem).
In the mid-nineteenth century, Pafnuty Chebyshev became the first person to think systematically in terms of the moments of random variables.[1]
The -th raw moment (i.e., moment about zero) of a random variable
X
f(x)
f(x)
c
It is possible to define moments for random variables in a more general fashion than moments for real-valued functions — see moments in metric spaces. The moment of a function, without further explanation, usually refers to the above expression with
c=0
Other moments may also be defined. For example, the th inverse moment about zero is
\operatorname{E}\left[X-n\right]
\operatorname{E}\left[lnn(X)\right].
The -th moment about zero of a probability density function
f(x)
Xn
\mu
If
f
Moment ordinal | Moment | Cumulant | ||||
---|---|---|---|---|---|---|
Raw | Central | Standardized | Raw | Normalized | ||
1 | 0 | 0 | Mean | |||
2 | – | 1 | Variance | 1 | ||
3 | – | – | – | Skewness | ||
4 | – | – | – | Excess kurtosis | ||
5 | – | – | Hyperskewness | – | – | |
6 | – | – | Hypertailedness | – | – | |
7+ | – | – | – | – | – |
See main article: Standardized moment. The normalised -th central moment or standardised moment is the -th central moment divided by ; the normalised -th central moment of the random variable is
These normalised central moments are dimensionless quantities, which represent the distribution independently of any linear change of scale.
See main article: Mean. The first raw moment is the mean, usually denoted
\mu\equiv\operatorname{E}[X].
\sigma\equiv\left(\operatorname{E}\left[(x-\mu)2\right]\right)
| ||||
.
See main article: Skewness. The third central moment is the measure of the lopsidedness of the distribution; any symmetric distribution will have a third central moment, if defined, of zero. The normalised third central moment is called the skewness, often . A distribution that is skewed to the left (the tail of the distribution is longer on the left) will have a negative skewness. A distribution that is skewed to the right (the tail of the distribution is longer on the right), will have a positive skewness.
For distributions that are not too different from the normal distribution, the median will be somewhere near ; the mode about .
See main article: Kurtosis.
The fourth central moment is a measure of the heaviness of the tail of the distribution. Since it is the expectation of a fourth power, the fourth central moment, where defined, is always nonnegative; and except for a point distribution, it is always strictly positive. The fourth central moment of a normal distribution is .
The kurtosis is defined to be the standardized fourth central moment. (Equivalently, as in the next section, excess kurtosis is the fourth cumulant divided by the square of the second cumulant.)[4] [5] If a distribution has heavy tails, the kurtosis will be high (sometimes called leptokurtic); conversely, light-tailed distributions (for example, bounded distributions such as the uniform) have low kurtosis (sometimes called platykurtic).
The kurtosis can be positive without limit, but must be greater than or equal to ; equality only holds for binary distributions. For unbounded skew distributions not too far from normal, tends to be somewhere in the area of and .
The inequality can be proven by consideringwhere . This is the expectation of a square, so it is non-negative for all a; however it is also a quadratic polynomial in a. Its discriminant must be non-positive, which gives the required relationship.
High-order moments are moments beyond 4th-order moments.
As with variance, skewness, and kurtosis, these are higher-order statistics, involving non-linear combinations of the data, and can be used for description or estimation of further shape parameters. The higher the moment, the harder it is to estimate, in the sense that larger samples are required in order to obtain estimates of similar quality. This is due to the excess degrees of freedom consumed by the higher orders. Further, they can be subtle to interpret, often being most easily understood in terms of lower order moments – compare the higher-order derivatives of jerk and jounce in physics. For example, just as the 4th-order moment (kurtosis) can be interpreted as "relative importance of tails as compared to shoulders in contribution to dispersion" (for a given amount of dispersion, higher kurtosis corresponds to thicker tails, while lower kurtosis corresponds to broader shoulders), the 5th-order moment can be interpreted as measuring "relative importance of tails as compared to center (mode and shoulders) in contribution to skewness" (for a given amount of skewness, higher 5th moment corresponds to higher skewness in the tail portions and little skewness of mode, while lower 5th moment corresponds to more skewness in shoulders).
Mixed moments are moments involving multiple variables.
The value
E[Xk]
k
k
X1...Xn
ki\geq0
k1 | |
E[{X | |
1} |
kn | |
… {X | |
n} |
]
k
k=k1+...+kn
E[(X1-E[X
k1 | |
1]) |
… (Xn-E[X
kn | |
n]) |
]
k
E[(X1-E[X1])(X2-E[X2])]
Some examples are covariance, coskewness and cokurtosis. While there is a unique covariance, there are multiple co-skewnesses and co-kurtoses.
Sincewhere is the binomial coefficient, it follows that the moments about b can be calculated from the moments about a by:
See main article: Convolution. The raw moment of a convolution reads where
\mun[ ⋅ ]
n
See main article: Cumulant.
The first raw moment and the second and third unnormalized central moments are additive in the sense that if X and Y are independent random variables then
(These can also hold for variables that satisfy weaker conditions than independence. The first always holds; if the second holds, the variables are called uncorrelated).
In fact, these are the first three cumulants and all cumulants share this additivity property.
For all k, the -th raw moment of a population can be estimated using the -th raw sample momentapplied to a sample drawn from the population.
It can be shown that the expected value of the raw sample moment is equal to the -th raw moment of the population, if that moment exists, for any sample size . It is thus an unbiased estimator. This contrasts with the situation for central moments, whose computation uses up a degree of freedom by using the sample mean. So for example an unbiased estimate of the population variance (the second central moment) is given byin which the previous denominator has been replaced by the degrees of freedom, and in which
\barX
\tfrac{n}{n-1},
See main article: Moment problem. Problems of determining a probability distribution from its sequence of moments are called problem of moments. Such problems were first discussed by P.L. Chebyshev (1874)[6] in connection with research on limit theorems. In order that the probability distribution of a random variable
X
\alphak=E\left[Xk\right]
{{\mun}':n=1,2,3,...}
\alphak(n)
k\geq1
\alphak
{\mun}'
\mu
\alphak
\mu
{\mun}'
\mu
Partial moments are sometimes referred to as "one-sided moments." The -th order lower and upper partial moments with respect to a reference point r may be expressed as
If the integral function do not converge, the partial moment does not exist.
Partial moments are normalized by being raised to the power 1/n. The upside potential ratio may be expressed as a ratio of a first-order upper partial moment to a normalized second-order lower partial moment.
Let be a metric space, and let B(M) be the Borel -algebra on M, the -algebra generated by the d-open subsets of M. (For technical reasons, it is also convenient to assume that M is a separable space with respect to the metric d.) Let .
The -th central moment of a measure on the measurable space (M, B(M)) about a given point is defined to be
μ is said to have finite -th central moment if the -th central moment of about x0 is finite for some .
This terminology for measures carries over to random variables in the usual way: if is a probability space and is a random variable, then the -th central moment of X about is defined to beand X has finite -th central moment if the -th central moment of X about x0 is finite for some .