In thermodynamics, the fundamental thermodynamic relation are four fundamental equations which demonstrate how four important thermodynamic quantities depend on variables that can be controlled and measured experimentally. Thus, they are essentially equations of state, and using the fundamental equations, experimental data can be used to determine sought-after quantities like G (Gibbs free energy) or H (enthalpy).[1] The relation is generally expressed as a microscopic change in internal energy in terms of microscopic changes in entropy, and volume for a closed system in thermal equilibrium in the following way.
dU=TdS-PdV
Here, U is internal energy, T is absolute temperature, S is entropy, P is pressure, and V is volume.
This is only one expression of the fundamental thermodynamic relation. It may be expressed in other ways, using different variables (e.g. using thermodynamic potentials). For example, the fundamental relation may be expressed in terms of the enthalpy H as
dH=TdS+VdP
in terms of the Helmholtz free energy F as
dF=-SdT-PdV
and in terms of the Gibbs free energy G as
dG=-SdT+VdP
The first law of thermodynamics states that:
dU=\deltaQ-\deltaW
where
\deltaQ
\deltaW
According to the second law of thermodynamics we have for a reversible process:
dS=
\deltaQ | |
T |
Hence:
\deltaQ=TdS
By substituting this into the first law, we have:
dU=TdS-\deltaW
Letting
\deltaW
\deltaW =PdV
we have:
dU=TdS-PdV
This equation has been derived in the case of reversible changes. However, since U, S, and V are thermodynamic state functions that depends on only the initial and final states of a thermodynamic process, the above relation holds also for non-reversible changes. If the composition, i.e. the amounts
ni
dU=TdS-PdV +\sumi\muidni
The
\mui
i
If the system has more external parameters than just the volume that can change, the fundamental thermodynamic relation generalizes to
dU=TdS+\sumjXjdxj+\sumi\muidni
Here the
Xj
xj
The fundamental thermodynamic relation and statistical mechanical principles can be derived from one another.
The above derivation uses the first and second laws of thermodynamics. The first law of thermodynamics is essentially a definition of heat, i.e. heat is the change in the internal energy of a system that is not caused by a change of the external parameters of the system.
However, the second law of thermodynamics is not a defining relation for the entropy. The fundamental definition of entropy of an isolated system containing an amount of energy
E
S=klog\left[\Omega\left(E\right)\right]
where
\Omega\left(E\right)
E
E+\deltaE
\deltaE
\deltaE
\deltaE
\deltaE
Deriving the fundamental thermodynamic relation from first principles thus amounts to proving that the above definition of entropy implies that for reversible processes we have:
dS=
\deltaQ | |
T |
The fundamental assumption of statistical mechanics is that all the
\Omega\left(E\right)
1 | \equiv\beta\equiv | |
kT |
dlog\left[\Omega\left(E\right)\right] | |
dE |
This definition can be derived from the microcanonical ensemble, which is a system of a constant number of particles, a constant volume and that does not exchange energy with its environment. Suppose that the system has some external parameter, x, that can be changed. In general, the energy eigenstates of the system will depend on x. According to the adiabatic theorem of quantum mechanics, in the limit of an infinitely slow change of the system's Hamiltonian, the system will stay in the same energy eigenstate and thus change its energy according to the change in energy of the energy eigenstate it is in.
The generalized force, X, corresponding to the external parameter x is defined such that
Xdx
Er
X=-
dEr | |
dx |
Since the system can be in any energy eigenstate within an interval of
\deltaE
X=-\left\langle
dEr | |
dx |
\right\rangle
To evaluate the average, we partition the
\Omega(E)
dEr | |
dx |
Y
Y+\deltaY
\OmegaY\left(E\right)
\Omega(E)=\sumY\OmegaY(E)
The average defining the generalized force can now be written:
X=-
1 | |
\Omega(E) |
\sumYY\OmegaY(E)
We can relate this to the derivative of the entropy with respect to x at constant energy E as follows. Suppose we change x to x + dx. Then
\Omega\left(E\right)
E
E+\deltaE
dEr | |
dx |
Y
Y+\deltaY
NY(E)=
\OmegaY(E) | |
\deltaE |
Ydx
such energy eigenstates. If
Ydx\leq\deltaE
E
E+\deltaE
\Omega
E+\deltaE
E+\deltaE
NY\left(E+\deltaE\right)
NY(E)-NY(E+\deltaE)
is thus the net contribution to the increase in
\Omega
\deltaE
E
E+\deltaE
NY(E)
NY(E+\deltaE)
Expressing the above expression as a derivative with respect to E and summing over Y yields the expression:
\left( | \partial\Omega |
\partialx |
\right)E=-\sumYY\left(
\partial\OmegaY | |
\partialE |
\right)x=\left(
\partial(\OmegaX) | |
\partialE |
\right)x
The logarithmic derivative of
\Omega
\left( | \partiallog\left(\Omega\right) |
\partialx |
\right)E=\betaX+\left(
\partialX | |
\partialE |
\right)x
The first term is intensive, i.e. it does not scale with system size. In contrast, the last term scales as the inverse system size and thus vanishes in the thermodynamic limit. We have thus found that:
\left( | \partialS |
\partialx |
\right)E=
X | |
T |
Combining this with
\left( | \partialS |
\partialE |
\right)x=
1 | |
T |
Gives:
dS=\left(
\partialS | |
\partialE |
\right)xdE+\left(
\partialS | |
\partialx |
\right)Edx=
dE | |
T |
+
X | |
T |
dx
which we can write as:
dE=TdS-Xdx
It has been shown that the fundamental thermodynamic relation together with the following three postulates[2] is sufficient to build the theory of statistical mechanics without the equal a priori probability postulate.
For example, in order to derive the Boltzmann distribution, we assume the probability density of microstate satisfies . The normalization factor (partition function) is therefore
Z=\sumif(Ei,T).
The entropy is therefore given by
S=kB\sumi
f(Ei,T) | log\left( | |
Z |
f(Ei,T) | |
Z |
\right).
If we change the temperature by while keeping the volume of the system constant, the change of entropy satisfies
dS=\left( | \partialS |
\partialT |
\right)VdT
where
\left( | \partialS |
\partialT |
\right)V=-kB
\sum | |||||||||||
|
=-kB\sumi
\partial | \left( | |
\partialT |
f(Ei,T) | |
Z |
\right) ⋅ logf(Ei,T)
Considering that
\left\langleE\right\rangle=\sumi
f(Ei,T) | |
Z |
⋅ Ei
we have
d\left\langleE\right\rangle=\sumi
\partial | \left( | |
\partialT |
f(Ei,T) | |
Z |
\right) ⋅ Ei ⋅ dT
From the fundamental thermodynamic relation, we have
- | dS | + |
kB |
d\left\langleE\right\rangle | |
kBT |
+
P | |
kBT |
dV=0
Since we kept constant when perturbing, we have . Combining the equations above, we have
\sumi
\partial | \left( | |
\partialT |
f(Ei,T) | |
Z |
\right) ⋅ \left[logf(Ei,T)+
Ei | |
kBT |
\right] ⋅ dT=0
Physics laws should be universal, i.e., the above equation must hold for arbitrary systems, and the only way for this to happen is
logf(Ei,T)+
Ei | |
kBT |
=0
That is
f(Ei,T)=\exp\left(-
Ei | |
kBT |
\right).
It has been shown that the third postulate in the above formalism can be replaced by the following:[3] However, the mathematical derivation will be much more complicated.