CFD Online Logo CFD Online URL
www.cfd-online.com
[Sponsors]
Home > Wiki > Finite difference

Finite difference

From CFD-Wiki

(Difference between revisions)
Jump to: navigation, search
(Transferred from wikipedia)
(References)
 
(5 intermediate revisions not shown)
Line 1: Line 1:
-
In [[mathematics]], a '''finite difference''' is like a differential quotient, except that it uses finite quantities instead of infinitesimal ones.
+
In mathematics, a '''finite difference''' is like a differential quotient, except that it uses finite quantities instead of infinitesimal ones.
-
The [[derivative]] of a function ''f'' at a point ''x'' is defined by the [[limit (mathematics)|limit]]
+
The derivative of a function ''f'' at a point ''x'' is defined by the limit
:<math> \lim_{h\to0} \frac{f(x+h) - f(x)}{h} </math>.
:<math> \lim_{h\to0} \frac{f(x+h) - f(x)}{h} </math>.
If ''h'' has a fixed (non-zero) value, instead of approaching zero, this quotient is called a ''finite difference''.
If ''h'' has a fixed (non-zero) value, instead of approaching zero, this quotient is called a ''finite difference''.
Line 7: Line 7:
==Calculus of finite differences==
==Calculus of finite differences==
-
One important aspect of finite differences is that it is analogous to the derivative. This means that [[difference operator]]s, mapping the function ''f'' to a finite difference, can be used to construct a [[calculus of finite differences]], which is similar to the [[differential calculus]] constructed from [[differential operator]]s.
+
One important aspect of finite differences is that it is analogous to the derivative. This means that difference operators, mapping the function ''f'' to a finite difference, can be used to construct a ''calculus of finite differences'', which is similar to the differential calculus constructed from differential operators.
==Numerical analysis==
==Numerical analysis==
-
Another important aspect is that finite differences approach differential quotients as ''h'' goes to zero. Thus, we can use finite differences to approximate derivatives. This is often used in [[numerical analysis]], especially in [[numerical ordinary differential equations]] and [[numerical partial differential equations]], which aim at the numerical solution of [[ordinary differential equation|ordinary]] and [[partial differential equation]]s respectively. The resulting methods are called ''finite-difference methods''.
+
Another important aspect is that finite differences approach differential quotients as ''h'' goes to zero. Thus, we can use finite differences to approximate derivatives. This is often used in [[Numerical methods|numerical analysis]], especially in numerical ordinary differential equations and numerical partial differential equations, which aim at the numerical solution of ordinary and partial differential equations respectively. The resulting methods are called ''finite-difference methods''.
For example, consider the ordinary differential equation
For example, consider the ordinary differential equation
Line 25: Line 25:
==  Example: the heat equation ==
==  Example: the heat equation ==
-
Consider the normalized [[heat equation]] in one dimension, with homogeneous [[Dirichlet boundary condition]]s
+
Consider the normalized [[heat equation]] in one dimension, with homogeneous [[Dirichlet boundary condition]]s:
:<math> U_t=U_{xx} \, </math>
:<math> U_t=U_{xx} \, </math>
Line 43: Line 43:
:<math> \frac{u_j^{n+1} - u_j^{n}}{k} =\frac{u_{j+1}^n - 2u_j^n + u_{j-1}^n}{h^2}. \, </math>
:<math> \frac{u_j^{n+1} - u_j^{n}}{k} =\frac{u_{j+1}^n - 2u_j^n + u_{j-1}^n}{h^2}. \, </math>
-
This is an [[explicit method]] for solving the one-dimensional [[heat equation]].
+
This is an explicit method for solving the one-dimensional heat equation.
We can obtain <math> u_j^{n+1} </math> from the other values this way:
We can obtain <math> u_j^{n+1} </math> from the other values this way:
Line 67: Line 67:
:<math> (1+2r)u_j^{n+1} - ru_{j-1}^{n+1} - ru_{j+1}^{n+1}= u_j^{n} </math>
:<math> (1+2r)u_j^{n+1} - ru_{j-1}^{n+1} - ru_{j+1}^{n+1}= u_j^{n} </math>
-
The scheme is always [[numerically stable]] and [[convergent]] but usually more numerically intensive then the explicit method as it requires solving a system of numerical equations on each time step. The errors are  linear over the time step and quadratic over the space step.
+
The scheme is always [[numerically stable]] and [[convergent]] but usually more numerically intensive than the explicit method as it requires solving a system of numerical equations on each time step. The errors are  linear over the time step and quadratic over the space step.
===Crank-Nicolson method===
===Crank-Nicolson method===
Finally if we use the central difference at time <math> t_{n+1/2} </math> and a second-order central difference for the space derivative at position <math> x_j </math> we get the recurrence equation:
Finally if we use the central difference at time <math> t_{n+1/2} </math> and a second-order central difference for the space derivative at position <math> x_j </math> we get the recurrence equation:
-
:<math> 2(\frac{u_j^{n+1} - u_j^{n}}{k}) =\frac{u_{j+1}^{n+1} - 2u_j^{n+1} + u_{j-1}^{n}}{h^2}+\frac{u_{j+1}^{n} - 2u_j^{n+1} + u_{j-1}^{n}}{h^2}. \, </math>
+
:<math> 2(\frac{u_j^{n+1} - u_j^{n}}{k}) =\frac{u_{j+1}^{n+1} - 2u_j^{n+1} + u_{j-1}^{n+1}}{h^2}+\frac{u_{j+1}^{n} - 2u_j^{n} + u_{j-1}^{n}}{h^2}. \, </math>
This formula is known as the [[Crank-Nicolson method]].
This formula is known as the [[Crank-Nicolson method]].
Line 78: Line 78:
We can obtain <math> u_j^{n+1} </math> from solving a system of linear equations:
We can obtain <math> u_j^{n+1} </math> from solving a system of linear equations:
-
:<math> (1+r)u_j^{n+1} - ru_{j-1}^{n+1} - ru_{j+1}^{n+1}= (1-r)u_j^n + ru_{j-1}^n + ru_{j+1}^n </math>
+
:<math> 2(1+r)u_j^{n+1} - ru_{j-1}^{n+1} - ru_{j+1}^{n+1}= 2(1-r)u_j^n + ru_{j-1}^n + ru_{j+1}^n </math>
The scheme is always [[numerically stable]] and [[convergent]] but usually more numerically intensive as it requires solving a system of numerical equations on each time step. The errors are quadratic over the time step and formally are of the fourth degree regarding the space step:
The scheme is always [[numerically stable]] and [[convergent]] but usually more numerically intensive as it requires solving a system of numerical equations on each time step. The errors are quadratic over the time step and formally are of the fourth degree regarding the space step:
Line 85: Line 85:
Usually the Crank-Nicolson scheme is the most accurate scheme for small time steps. The explicit scheme is the least accurate and can be unstable, but is also the easiest to implement and the least numerically intensive. The implicit scheme works the best for large time steps.
Usually the Crank-Nicolson scheme is the most accurate scheme for small time steps. The explicit scheme is the least accurate and can be unstable, but is also the easiest to implement and the least numerically intensive. The implicit scheme works the best for large time steps.
-
 
-
==See also==
 
-
* [[Newton series]]
 
-
* [[Binomial transform]]
 
-
* [[Divided differences]]
 
-
* [[Finite element method]]
 
-
* [[Multigrid method]]
 
==References==
==References==
Line 97: Line 90:
* Oliver Rübenkönig, ''[http://www.imtek.uni-freiburg.de/simulation/mathematica/IMSweb/imsTOC/Lectures%20and%20Tips/Simulation%20I/FDM_introDocu.html The Finite Difference Method (FDM) - An introduction]'', (2006) [[Albert Ludwigs University of Freiburg]]
* Oliver Rübenkönig, ''[http://www.imtek.uni-freiburg.de/simulation/mathematica/IMSweb/imsTOC/Lectures%20and%20Tips/Simulation%20I/FDM_introDocu.html The Finite Difference Method (FDM) - An introduction]'', (2006) [[Albert Ludwigs University of Freiburg]]
*[http://en.wikipedia.org/wiki/Finite_difference Finite Difference article on Wikipedia]
*[http://en.wikipedia.org/wiki/Finite_difference Finite Difference article on Wikipedia]
 +
* H. Lomax, T.H. Pulliam and D.W. Zingg, ''Fundamentals of Computational Fluid Dynamics'', Series: Scientific Computation , Springer-Verlag, 2001, ISBN: 3-540-41607-2

Latest revision as of 16:29, 13 May 2008

In mathematics, a finite difference is like a differential quotient, except that it uses finite quantities instead of infinitesimal ones.

The derivative of a function f at a point x is defined by the limit

 \lim_{h\to0} \frac{f(x+h) - f(x)}{h} .

If h has a fixed (non-zero) value, instead of approaching zero, this quotient is called a finite difference.

Contents

Calculus of finite differences

One important aspect of finite differences is that it is analogous to the derivative. This means that difference operators, mapping the function f to a finite difference, can be used to construct a calculus of finite differences, which is similar to the differential calculus constructed from differential operators.

Numerical analysis

Another important aspect is that finite differences approach differential quotients as h goes to zero. Thus, we can use finite differences to approximate derivatives. This is often used in numerical analysis, especially in numerical ordinary differential equations and numerical partial differential equations, which aim at the numerical solution of ordinary and partial differential equations respectively. The resulting methods are called finite-difference methods.

For example, consider the ordinary differential equation

 u'(x) = 3u(x) + 2. \,

The Euler method for solving this equation uses the finite difference

\frac{u(x+h) - u(x)}{h} \approx u'(x)

to approximate the differential equation by

 u(x+h) = u(x) + h(3u(x)+2). \,

The last equation is called a finite-difference equation. Solving this equation gives an approximate solution to the differential equation.

The error between the approximate solution and the true solution is determined by the error that is made by going from a differential operator to a difference operator. This error is called the discretization error or truncation error (the term truncation error reflects the fact that a difference operator can be viewed as a finite part of the infinite Taylor series of the differential operator).

Example: the heat equation

Consider the normalized heat equation in one dimension, with homogeneous Dirichlet boundary conditions:

 U_t=U_{xx} \,
 U(0,t)=U(1,t)=0 \, (boundary condition)
 U(x,0) =U_0(x) \, (initial condition)

One way to numerically solve this equation is to approximate all the derivatives by finite differences. We partition the domain in space using a mesh  x_0, ..., x_J and in time using a mesh  t_0, ...., t_N . We assume a uniform partition both in space and in time, so the difference between two consecutive space points will be h and between two consecutive time points will be k. The points

 u(x_j,t_n) = u_j^n

will represent the numerical approximation of  U(x_j, t_n).

Explicit method

Using a forward difference at time  t_n and a second-order central difference for the space derivative at position  x_j we get the recurrence equation:

 \frac{u_j^{n+1} - u_j^{n}}{k} =\frac{u_{j+1}^n - 2u_j^n + u_{j-1}^n}{h^2}. \,

This is an explicit method for solving the one-dimensional heat equation.

We can obtain  u_j^{n+1} from the other values this way:

 u_j^{n+1} = (1-2r)u_j^n + ru_{j-1}^n + ru_{j+1}^n

where  r=k/h^2.

So, knowing the values at time n you can obtain the corresponding ones at time n+1 using this recurrence relation.  u_0^n and  u_J^n must be replaced by the border conditions, in this example they are both 0.

This explicit method is known to be numerically stable and convergent whenever  r\le 1/2. . The numerical errors are proportional to the time step and the square of the space step:

 \Delta u = O(k)+O(h^2)  \,

Implicit method

If we use the backward difference at time  t_{n+1} and a second-order central difference for the space derivative at position  x_j we get the recurrence equation:

 \frac{u_j^{n+1} - u_j^{n}}{k} =\frac{u_{j+1}^{n+1} - 2u_j^{n+1} + u_{j-1}^{n+1}}{h^2}. \,

This is an implicit method for solving the one-dimensional heat equation.

We can obtain  u_j^{n+1} from solving a system of linear equations:

 (1+2r)u_j^{n+1} - ru_{j-1}^{n+1} - ru_{j+1}^{n+1}= u_j^{n}

The scheme is always numerically stable and convergent but usually more numerically intensive than the explicit method as it requires solving a system of numerical equations on each time step. The errors are linear over the time step and quadratic over the space step.

Crank-Nicolson method

Finally if we use the central difference at time  t_{n+1/2} and a second-order central difference for the space derivative at position  x_j we get the recurrence equation:

 2(\frac{u_j^{n+1} - u_j^{n}}{k}) =\frac{u_{j+1}^{n+1} - 2u_j^{n+1} + u_{j-1}^{n+1}}{h^2}+\frac{u_{j+1}^{n} - 2u_j^{n} + u_{j-1}^{n}}{h^2}. \,

This formula is known as the Crank-Nicolson method.

We can obtain  u_j^{n+1} from solving a system of linear equations:

 2(1+r)u_j^{n+1} - ru_{j-1}^{n+1} - ru_{j+1}^{n+1}= 2(1-r)u_j^n + ru_{j-1}^n + ru_{j+1}^n

The scheme is always numerically stable and convergent but usually more numerically intensive as it requires solving a system of numerical equations on each time step. The errors are quadratic over the time step and formally are of the fourth degree regarding the space step:

 \Delta u = O(k^2)+O(h^4)  \,

(in reality, usually near the boundaries, the scheme failed to provide the fourth order accuracy over the space and degenerates to the quadratic errors).

Usually the Crank-Nicolson scheme is the most accurate scheme for small time steps. The explicit scheme is the least accurate and can be unstable, but is also the easiest to implement and the least numerically intensive. The implicit scheme works the best for large time steps.

References

My wiki