In numerical analysis, finite-difference methods (FDM) are a class of numerical techniques for solving differential equations by approximating derivatives with finite differences. Both the spatial domain and time domain (if applicable) are discretized, or broken into a finite number of intervals, and the values of the solution at the end points of the intervals are approximated by solving algebraic equations containing finite differences and values from nearby points.
Finite difference methods convert ordinary differential equations (ODE) or partial differential equations (PDE), which may be nonlinear, into a system of linear equations that can be solved by matrix algebra techniques. Modern computers can perform these linear algebra computations efficiently which, along with their relative ease of implementation, has led to the widespread use of FDM in modern numerical analysis.Today, FDMs are one of the most common approaches to the numerical solution of PDE, along with finite element methods.[1]
For a n-times differentiable function, by Taylor's theorem the Taylor series expansion is given as
Where n! denotes the factorial of n, and Rn(x) is a remainder term, denoting the difference between the Taylor polynomial of degree n and the original function.
Following is the process to derive an approximation for the first derivative of the function f by first truncating the Taylor polynomial plus remainder:Dividing across by h gives:Solving for
f'(x0)
Assuming that
R1(x)
This is similar to the definition of derivative, which is:except for the limit towards zero (the method is named after this).
See also: Finite difference coefficient.
The error in a method's solution is defined as the difference between the approximation and the exact analytical solution. The two sources of error in finite difference methods are round-off error, the loss of precision due to computer rounding of decimal quantities, and truncation error or discretization error, the difference between the exact solution of the original differential equation and the exact quantity assuming perfect arithmetic (no round-off).
To use a finite difference method to approximate the solution to a problem, one must first discretize the problem's domain. This is usually done by dividing the domain into a uniform grid (see image). This means that finite-difference methods produce sets of discrete numerical approximations to the derivative, often in a "time-stepping" manner.
An expression of general interest is the local truncation error of a method. Typically expressed using Big-O notation, local truncation error refers to the error from a single application of a method. That is, it is the quantity
f'(xi)-f'i
f'(xi)
f'i
f(x0+h)
f(xi)=f(x0+ih)
In this case, the local truncation error is proportional to the step sizes. The quality and duration of simulated FDM solution depends on the discretization equation selection and the step sizes (time and space steps). The data quality and simulation duration increase significantly with smaller step size.[2] Therefore, a reasonable balance between data quality and simulation duration is necessary for practical usage. Large time steps are useful for increasing simulation speed in practice. However, time steps which are too large may create instabilities and affect the data quality.[3] [4]
The von Neumann and Courant-Friedrichs-Lewy criteria are often evaluated to determine the numerical model stability.[5] [6]
For example, consider the ordinary differential equationThe Euler method for solving this equation uses the finite difference quotientto approximate the differential equation by first substituting it for u'(x) then applying a little algebra (multiplying both sides by h, and then adding u(x) to both sides) to getThe last equation is a finite-difference equation, and solving this equation gives an approximate solution to the differential equation.
Consider the normalized heat equation in one dimension, with homogeneous Dirichlet boundary conditions
One way to numerically solve this equation is to approximate all the derivatives by finite differences. First partition the domain in space using a mesh
x0,...,xJ
t0,...,tN
will represent the numerical approximation of
u(xj,tn).
Using a forward difference at time
tn
xj
This is an explicit method for solving the one-dimensional heat equation.
One can obtain
n+1 | |
u | |
j |
where
r=k/h2.
So, with this recurrence relation, and knowing the values at time n, one can obtain the corresponding values at time n+1.
n | |
u | |
0 |
n | |
u | |
J |
This explicit method is known to be numerically stable and convergent whenever
r\le1/2
Using the backward difference at time
tn+1
xj
This is an implicit method for solving the one-dimensional heat equation.
One can obtain
n+1 | |
u | |
j |
The scheme is always numerically stable and convergent but usually more numerically intensive than the explicit method as it requires solving a system of numerical equations on each time step. The errors are linear over the time step and quadratic over the space step:
Finally, using the central difference at time
tn+1/2
xj
This formula is known as the Crank - Nicolson method.
One can obtain
n+1 | |
u | |
j |
The scheme is always numerically stable and convergent but usually more numerically intensive as it requires solving a system of numerical equations on each time step. The errors are quadratic over both the time step and the space step:
To summarize, usually the Crank - Nicolson scheme is the most accurate scheme for small time steps. For larger time steps, the implicit scheme works better since it is less computationally demanding. The explicit scheme is the least accurate and can be unstable, but is also the easiest to implement and the least numerically intensive.
Here is an example. The figures below present the solutions given by the above methods to approximate the heat equation
with the boundary condition
The exact solution is
The (continuous) Laplace operator in
n
\Deltau(x)=
n | |
\sum | |
i=1 |
2 | |
\partial | |
i |
u(x)
\Deltahu
n
In 1D the Laplace operator is approximated as This approximation is usually expressed via the following stenciland which represents a symmetric, tridiagonal matrix. For an equidistant grid one gets a Toeplitz matrix.
The 2D case shows all the characteristics of the more general n-dimensional case. Each second partial derivative needs to be approximated similar to the 1D case which is usually given by the following stencil
Consistency of the above-mentioned approximation can be shown for highly regular functions, such as
u\inC4(\Omega)
To prove this, one needs to substitute Taylor Series expansions up to order 3 into the discrete Laplace operator.
Similar to continuous subharmonic functions one can define subharmonic functions for finite-difference approximations
uh
One can define a general stencil of positive type via
If
uh
A similar mean value property also holds for the continuous case.
For a (discrete) subharmonic function
uh
\Omegah,\partial\Omegah
\Omega
\partial\Omega
A similar maximum principle also holds for the continuous case.
The SBP-SAT (summation by parts - simultaneous approximation term) method is a stable and accurate technique for discretizing and imposing boundary conditions of a well-posed partial differential equation using high order finite differences.[8] [9]
The method is based on finite differences where the differentiation operators exhibit summation-by-parts properties. Typically, these operators consist of differentiation matrices with central difference stencils in the interior with carefully chosen one-sided boundary stencils designed to mimic integration-by-parts in the discrete setting. Using the SAT technique, the boundary conditions of the PDE are imposed weakly, where the boundary values are "pulled" towards the desired conditions rather than exactly fulfilled. If the tuning parameters (inherent to the SAT technique) are chosen properly, the resulting system of ODE's will exhibit similar energy behavior as the continuous PDE, i.e. the system has no non-physical energy growth. This guarantees stability if an integration scheme with a stability region that includes parts of the imaginary axis, such as the fourth order Runge-Kutta method, is used. This makes the SAT technique an attractive method of imposing boundary conditions for higher order finite difference methods, in contrast to for example the injection method, which typically will not be stable if high order differentiation operators are used.