In physics and probability theory, Mean-field theory (MFT) or Self-consistent field theory studies the behavior of high-dimensional random (stochastic) models by studying a simpler model that approximates the original by averaging over degrees of freedom (the number of values in the final calculation of a statistic that are free to vary). Such models consider many individual components that interact with each other.
The main idea of MFT is to replace all interactions to any one body with an average or effective interaction, sometimes called a molecular field.[1] This reduces any many-body problem into an effective one-body problem. The ease of solving MFT problems means that some insight into the behavior of the system can be obtained at a lower computational cost.
MFT has since been applied to a wide range of fields outside of physics, including statistical inference, graphical models, neuroscience,[2] artificial intelligence, epidemic models,[3] queueing theory,[4] computer-network performance and game theory,[5] as in the quantal response equilibrium.
The idea first appeared in physics (statistical mechanics) in the work of Pierre Curie[6] and Pierre Weiss to describe phase transitions.[7] MFT has been used in the Bragg–Williams approximation, models on Bethe lattice, Landau theory, Pierre–Weiss approximation, Flory–Huggins solution theory, and Scheutjens–Fleer theory.
Systems with many (sometimes infinite) degrees of freedom are generally hard to solve exactly or compute in closed, analytic form, except for some simple cases (e.g. certain Gaussian random-field theories, the 1D Ising model). Often combinatorial problems arise that make things like computing the partition function of a system difficult. MFT is an approximation method that often makes the original problem to be solvable and open to calculation, and in some cases MFT may give very accurate approximations.
In field theory, the Hamiltonian may be expanded in terms of the magnitude of fluctuations around the mean of the field. In this context, MFT can be viewed as the "zeroth-order" expansion of the Hamiltonian in fluctuations. Physically, this means that an MFT system has no fluctuations, but this coincides with the idea that one is replacing all interactions with a "mean-field”.
Quite often, MFT provides a convenient launch point for studying higher-order fluctuations. For example, when computing the partition function, studying the combinatorics of the interaction terms in the Hamiltonian can sometimes at best produce perturbation results or Feynman diagrams that correct the mean-field approximation.
In general, dimensionality plays an active role in determining whether a mean-field approach will work for any particular problem. There is sometimes a critical dimension above which MFT is valid and below which it is not.
Heuristically, many interactions are replaced in MFT by one effective interaction. So if the field or particle exhibits many random interactions in the original system, they tend to cancel each other out, so the mean effective interaction and MFT will be more accurate. This is true in cases of high dimensionality, when the Hamiltonian includes long-range forces, or when the particles are extended (e.g. polymers). The Ginzburg criterion is the formal expression of how fluctuations render MFT a poor approximation, often depending upon the number of spatial dimensions in the system of interest.
The formal basis for mean-field theory is the Bogoliubov inequality. This inequality states that the free energy of a system with Hamiltonian
l{H}=l{H}0+\Deltal{H}
has the following upper bound:
F\leqF0 \stackrel{def
where
S0
F
F0
l{H}0
l{H}0=
N | |
\sum | |
i=1 |
hi(\xii),
where
\xii
For the most common case that the target Hamiltonian contains only pairwise interactions, i.e.,
l{H}=\sum(i,j)
where
l{P}
\operatorname{Tr}if(\xii)
f
\begin{align} F0&=\operatorname{Tr}1,2,\ldots,Nl{H}(\xi1,\xi2,\ldots,\xiN)
(N) | |
P | |
0(\xi |
1,\xi2,\ldots,\xiN)\\ &+kT\operatorname{Tr}1,2,\ldots,N
(N) | |
P | |
0(\xi |
1,\xi2,\ldots,\xiN)log
(N) | |
P | |
0(\xi |
1,\xi2,\ldots,\xiN), \end{align}
where
(N) | |
P | |
0(\xi |
1,\xi2,...,\xiN)
(\xi1,\xi2,...,\xiN)
\begin{align}
(N) | |
P | |
0(\xi |
1,\xi2,\ldots,\xiN) &=
1 | ||||||
|
-\betal{H | |
e | |
0(\xi |
1,\xi2,\ldots,\xiN)}\\ &=
N | |
\prod | |
i=1 |
1 | |
Z0 |
-\betahi(\xii) | |
e |
\stackrel{def
where
Z0
\begin{align} F0&=\sum(i,j)
In order to minimise, we take the derivative with respect to the single-degree-of-freedom probabilities
(i) | |
P | |
0 |
(i) | |
P | |
0(\xi |
i)=
1 | |
Z0 |
| |||||||||
e |
, i=1,2,\ldots,N,
where the mean field is given by
MF(\xi | |
h | |
i) |
=\sum\{j\}}\operatorname{Tr}jVi,j(\xii,\xij)
(j) | |
P | |
0(\xi |
j).
Mean field theory can be applied to a number of physical systems so as to study phenomena such as phase transitions.[8]
The Bogoliubov inequality, shown above, can be used to find the dynamics of a mean field model of the two-dimensional Ising lattice. A magnetisation function can be calculated from the resultant approximate free energy.[9] The first step is choosing a more tractable approximation of the true Hamiltonian. Using a non-interacting or effective field Hamiltonian,
-m\sumisi
the variational free energy is
FV=F0+\left\langle\left(-J\sumsisj-h\sumsi\right)-\left(-m\sumsi\right)\right\rangle0.
By the Bogoliubov inequality, simplifying this quantity and calculating the magnetisation function that minimises the variational free energy yields the best approximation to the actual magnetisation. The minimiser is
m=J\sum\langlesj\rangle0+h,
m=tanh(zJ\betam)+h.
Equating the effective field felt by all spins to a mean spin value relates the variational approach to the suppression of fluctuations. The physical interpretation of the magnetisation function is then a field of mean values for individual spins.
Consider the Ising model on a
d
H=-J\sum\langlesisj-h\sumisi,
\sum\langle
\langlei,j\rangle
si,sj=\pm1
Let us transform our spin variable by introducing the fluctuation from its mean value
mi\equiv\langlesi\rangle
H=-J\sum\langle(mi+\deltasi)(mj+\deltasj)-h\sumisi,
where we define
\deltasi\equivsi-mi
If we expand the right side, we obtain one term that is entirely dependent on the mean values of the spins and independent of the spin configurations. This is the trivial term, which does not affect the statistical properties of the system. The next term is the one involving the product of the mean value of the spin and the fluctuation value. Finally, the last term involves a product of two fluctuation values.
The mean field approximation consists of neglecting this second-order fluctuation term:
H ≈ HMF\equiv-J\sum\langle(mimj+mi\deltasj+mj\deltasi)-h\sumisi.
These fluctuations are enhanced at low dimensions, making MFT a better approximation for high dimensions.
Again, the summand can be re-expanded. In addition, we expect that the mean value of each spin is site-independent, since the Ising chain is translationally invariant. This yields
HMF=-J\sum\langle(m2+2m(si-m))-h\sumisi.
The summation over neighboring spins can be rewritten as
\sum\langle=
1 | |
2 |
\sumi\sumj
nn(i)
i
1/2
HMF=
Jm2Nz | |
2 |
-\underbrace{(h+mJ
z)} | |
heff. |
\sumisi,
where
z
heff.=h+Jzm
h
d
z=2d
Substituting this Hamiltonian into the partition function and solving the effective 1D problem, we obtain
Z=
| ||||
e |
\left[2\cosh\left(
h+mJz | |
kBT |
\right)\right]N,
where
N
m
heff.
We thus have two equations between
m
heff.
m
Tc
m=0
T<Tc
m=\pmm0
Tc
Tc=
Jz | |
kB |
This shows that MFT can account for the ferromagnetic phase transition.
Similarly, MFT can be applied to other types of Hamiltonian as in the following cases:
\Delta
Variationally minimisation like mean field theory can be also be used in statistical inference.
See main article: Dynamical mean field theory.
In mean field theory, the mean field appearing in the single-site problem is a time-independent scalar or vector quantity. However, this isn't always the case: in a variant of mean field theory called dynamical mean field theory (DMFT), the mean field becomes a time-dependent quantity. For instance, DMFT can be applied to the Hubbard model to study the metal–Mott-insulator transition.