Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Finite difference method

From Wikipedia, the free encyclopedia
Class of numerical techniques
Differential equations
Scope
Classification
Solution
People

Innumerical analysis,finite-difference methods (FDM) are a class of numerical techniques for solvingdifferential equations by approximatingderivatives withfinite differences. Both the spatial domain and time domain (if applicable) arediscretized, or broken into a finite number of intervals, and the values of the solution at the end points of the intervals are approximated by solving algebraic equations containing finite differences and values from nearby points.

Finite difference methods convertordinary differential equations (ODE) orpartial differential equations (PDE), which may benonlinear, into asystem of linear equations that can be solved bymatrix algebra techniques. Modern computers can perform theselinear algebra computations efficiently, and this, along with their relative ease of implementation, has led to the widespread use of FDM in modern numerical analysis.[1]Today, FDMs are one of the most common approaches to the numerical solution of PDE, along withfinite element methods.[1]

Derive difference quotient from Taylor's polynomial

[edit]

For an-times differentiable function, byTaylor's theorem theTaylor series expansion is given asf(x0+h)=f(x0)+f(x0)1!h+f(2)(x0)2!h2++f(n)(x0)n!hn+Rn(x),{\displaystyle f(x_{0}+h)=f(x_{0})+{\frac {f'(x_{0})}{1!}}h+{\frac {f^{(2)}(x_{0})}{2!}}h^{2}+\cdots +{\frac {f^{(n)}(x_{0})}{n!}}h^{n}+R_{n}(x),}

Wheren! denotes thefactorial ofn, andRn(x) is a remainder term, denoting the difference between the Taylor polynomial of degreen and the original function.

Following is the process to derive an approximation for the first derivative of the functionf by first truncating the Taylor polynomial plus remainder:f(x0+h)=f(x0)+f(x0)h+R1(x).{\displaystyle f(x_{0}+h)=f(x_{0})+f'(x_{0})h+R_{1}(x).}Dividing across byh gives:f(x0+h)h=f(x0)h+f(x0)+R1(x)h{\displaystyle {f(x_{0}+h) \over h}={f(x_{0}) \over h}+f'(x_{0})+{R_{1}(x) \over h}}Solving forf(x0){\displaystyle f'(x_{0})}:f(x0)=f(x0+h)f(x0)hR1(x)h.{\displaystyle f'(x_{0})={f(x_{0}+h)-f(x_{0}) \over h}-{R_{1}(x) \over h}.}

Assuming thatR1(x){\displaystyle R_{1}(x)} is sufficiently small, the approximation of the first derivative off is:f(x0)f(x0+h)f(x0)h.{\displaystyle f'(x_{0})\approx {f(x_{0}+h)-f(x_{0}) \over h}.}

This is similar to the definition of derivative, which is:f(x0)=limh0f(x0+h)f(x0)h.{\displaystyle f'(x_{0})=\lim _{h\to 0}{\frac {f(x_{0}+h)-f(x_{0})}{h}}.}except for the limit towards zero (the method is named after this).

Accuracy and order

[edit]
See also:Finite difference coefficient

The error in a method's solution is defined as the difference between the approximation and the exact analytical solution. The two sources of error in finite difference methods areround-off error, the loss of precision due to computer rounding of decimal quantities, andtruncation error ordiscretization error, the difference between the exact solution of the original differential equation and the exact quantity assuming perfect arithmetic (no round-off).

The finite difference method relies on discretizing a function on a grid.

To use a finite difference method to approximate the solution to a problem, one must first discretize the problem's domain. This is usually done by dividing the domain into a uniform grid (see image). This means that finite-difference methods produce sets of discrete numerical approximations to the derivative, often in a "time-stepping" manner.

An expression of general interest is thelocal truncation error of a method. Typically expressed usingBig-O notation, local truncation error refers to the error from a single application of a method. That is, it is the quantityf(xi)fi{\displaystyle f'(x_{i})-f'_{i}} iff(xi){\displaystyle f'(x_{i})} refers to the exact value andfi{\displaystyle f'_{i}} to the numerical approximation. The remainder term of the Taylor polynomial can be used to analyzelocal truncation error. Using theLagrange form of the remainder from the Taylor polynomial forf(x0+h){\displaystyle f(x_{0}+h)}, which isRn(x0+h)=f(n+1)(ξ)(n+1)!(h)n+1,x0<ξ<x0+h,{\displaystyle R_{n}(x_{0}+h)={\frac {f^{(n+1)}(\xi )}{(n+1)!}}(h)^{n+1}\,,\quad x_{0}<\xi <x_{0}+h,}the dominant term of the local truncation error can be discovered. For example, again using the forward-difference formula for the first derivative, knowing thatf(xi)=f(x0+ih){\displaystyle f(x_{i})=f(x_{0}+ih)},f(x0+ih)=f(x0)+f(x0)ih+f(ξ)2!(ih)2,{\displaystyle f(x_{0}+ih)=f(x_{0})+f'(x_{0})ih+{\frac {f''(\xi )}{2!}}(ih)^{2},}and with some algebraic manipulation, this leads tof(x0+ih)f(x0)ih=f(x0)+f(ξ)2!ih,{\displaystyle {\frac {f(x_{0}+ih)-f(x_{0})}{ih}}=f'(x_{0})+{\frac {f''(\xi )}{2!}}ih,}and further noting that the quantity on the left is the approximation from the finite difference method and that the quantity on the right is the exact quantity of interest plus a remainder, clearly that remainder is the local truncation error. A final expression of this example and its order is:f(x0+ih)f(x0)ih=f(x0)+O(h).{\displaystyle {\frac {f(x_{0}+ih)-f(x_{0})}{ih}}=f'(x_{0})+O(h).}

In this case, the local truncation error is proportional to the step sizes. The quality and duration of simulated FDM solution depends on the discretization equation selection and the step sizes (time and space steps). The data quality and simulation duration increase significantly with smaller step size.[2] Therefore, a reasonable balance between data quality and simulation duration is necessary for practical usage. Large time steps are useful for increasing simulation speed in practice. However, time steps which are too large may create instabilities and affect the data quality.[3][4]

Thevon Neumann andCourant-Friedrichs-Lewy criteria are often evaluated to determine the numerical model stability.[3][4][5][6]

Example: ordinary differential equation

[edit]

For example, consider the ordinary differential equationu(x)=3u(x)+2.{\displaystyle u'(x)=3u(x)+2.}TheEuler method for solving this equation uses the finite difference quotientu(x+h)u(x)hu(x){\displaystyle {\frac {u(x+h)-u(x)}{h}}\approx u'(x)}to approximate the differential equation by first substituting it foru(x){\displaystyle u'(x)} then applying a little algebra (multiplying both sides byh, and then addingu(x){\displaystyle u(x)} to both sides) to getu(x+h)u(x)+h(3u(x)+2).{\displaystyle u(x+h)\approx u(x)+h(3u(x)+2).}The last equation is a finite-difference equation, and solving this equation gives an approximate solution to the differential equation.

Example: The heat equation

[edit]

Consider the normalizedheat equation in one dimension, with homogeneousDirichlet boundary conditions

{Ut=UxxU(0,t)=U(1,t)=0(boundary condition)U(x,0)=U0(x)(initial condition){\displaystyle {\begin{cases}U_{t}=U_{xx}\\U(0,t)=U(1,t)=0&{\text{(boundary condition)}}\\U(x,0)=U_{0}(x)&{\text{(initial condition)}}\end{cases}}}

One way to numerically solve this equation is to approximate all the derivatives by finite differences. First partition the domain in space using a meshx0,,xJ{\displaystyle x_{0},\dots ,x_{J}} and in time using a mesht0,,tN{\displaystyle t_{0},\dots ,t_{N}}. Assume a uniform partition both in space and in time, so the difference between two consecutive space points will beh and between two consecutive time points will bek. The points

u(xj,tn)=ujn{\displaystyle u(x_{j},t_{n})=u_{j}^{n}}

will represent the numerical approximation ofu(xj,tn).{\displaystyle u(x_{j},t_{n}).}

Explicit method

[edit]
Thestencil for the most common explicit method for the heat equation.

Using aforward differenceat timetn{\displaystyle t_{n}} and a second-ordercentral difference for the space derivative at positionxj{\displaystyle x_{j}} (FTCS) gives the recurrence equation:

ujn+1ujnk=uj+1n2ujn+uj1nh2.{\displaystyle {\frac {u_{j}^{n+1}-u_{j}^{n}}{k}}={\frac {u_{j+1}^{n}-2u_{j}^{n}+u_{j-1}^{n}}{h^{2}}}.}

This is anexplicit method for solving the one-dimensionalheat equation.

One can obtainujn+1{\displaystyle u_{j}^{n+1}} from the other values this way:

ujn+1=(12r)ujn+ruj1n+ruj+1n{\displaystyle u_{j}^{n+1}=(1-2r)u_{j}^{n}+ru_{j-1}^{n}+ru_{j+1}^{n}}

wherer=k/h2.{\displaystyle r=k/h^{2}.}

So, with this recurrence relation, and knowing the values at timen, one can obtain the corresponding values at timen+1.u0n{\displaystyle u_{0}^{n}} anduJn{\displaystyle u_{J}^{n}} must be replaced by the boundary conditions, in this example they are both 0.

This explicit method is known to benumerically stable andconvergent wheneverr1/2{\displaystyle r\leq 1/2}.[7] The numerical errors are proportional to the time step and the square of the space step:Δu=O(k)+O(h2){\displaystyle \Delta u=O(k)+O(h^{2})}

Implicit method

[edit]
The implicit method stencil.

Using thebackward differenceat timetn+1{\displaystyle t_{n+1}} and a second-order central difference for the space derivative at positionxj{\displaystyle x_{j}} (The Backward Time, Centered Space Method "BTCS") gives the recurrence equation:

ujn+1ujnk=uj+1n+12ujn+1+uj1n+1h2.{\displaystyle {\frac {u_{j}^{n+1}-u_{j}^{n}}{k}}={\frac {u_{j+1}^{n+1}-2u_{j}^{n+1}+u_{j-1}^{n+1}}{h^{2}}}.}

This is animplicit method for solving the one-dimensionalheat equation.

One can obtainujn+1{\displaystyle u_{j}^{n+1}} from solving a system of linear equations:

(1+2r)ujn+1ruj1n+1ruj+1n+1=ujn{\displaystyle (1+2r)u_{j}^{n+1}-ru_{j-1}^{n+1}-ru_{j+1}^{n+1}=u_{j}^{n}}

The scheme is alwaysnumerically stable and convergent but usually more numerically intensive than the explicit method as it requires solving a system of numerical equations on each time step. The errors are linear over the time step and quadratic over the space step:Δu=O(k)+O(h2).{\displaystyle \Delta u=O(k)+O(h^{2}).}

Crank–Nicolson method

[edit]

Finally, using the central difference at timetn+1/2{\displaystyle t_{n+1/2}} and a second-order central difference for the space derivative at positionxj{\displaystyle x_{j}} ("CTCS") gives the recurrence equation:

ujn+1ujnk=12(uj+1n+12ujn+1+uj1n+1h2+uj+1n2ujn+uj1nh2).{\displaystyle {\frac {u_{j}^{n+1}-u_{j}^{n}}{k}}={\frac {1}{2}}\left({\frac {u_{j+1}^{n+1}-2u_{j}^{n+1}+u_{j-1}^{n+1}}{h^{2}}}+{\frac {u_{j+1}^{n}-2u_{j}^{n}+u_{j-1}^{n}}{h^{2}}}\right).}

This formula is known as theCrank–Nicolson method.

The Crank–Nicolson stencil.

One can obtainujn+1{\displaystyle u_{j}^{n+1}} from solving a system of linear equations:

(2+2r)ujn+1ruj1n+1ruj+1n+1=(22r)ujn+ruj1n+ruj+1n{\displaystyle (2+2r)u_{j}^{n+1}-ru_{j-1}^{n+1}-ru_{j+1}^{n+1}=(2-2r)u_{j}^{n}+ru_{j-1}^{n}+ru_{j+1}^{n}}

The scheme is alwaysnumerically stable and convergent but usually more numerically intensive as it requires solving a system of numerical equations on each time step. The errors are quadratic over both the time step and the space step:Δu=O(k2)+O(h2).{\displaystyle \Delta u=O(k^{2})+O(h^{2}).}

Comparison

[edit]

To summarize, usually theCrank–Nicolson scheme is the most accurate scheme for small time steps. For larger time steps, the implicit scheme works better since it is less computationally demanding. The explicit scheme is the least accurate and can be unstable, but is also the easiest to implement and the least numerically intensive.

Here is an example. The figures below present the solutions given by the above methods to approximate the heat equation

Ut=αUxx,α=1π2,{\displaystyle U_{t}=\alpha U_{xx},\quad \alpha ={\frac {1}{\pi ^{2}}},}

with the boundary condition

U(0,t)=U(1,t)=0.{\displaystyle U(0,t)=U(1,t)=0.}

The exact solution is

U(x,t)=1π2etsin(πx).{\displaystyle U(x,t)={\frac {1}{\pi ^{2}}}e^{-t}\sin(\pi x).}

Comparison of Finite Difference Methods
c = 4
Explicit method (not stable)
c = 6
Implicit method (stable)
c = 8.5
Crank-Nicolson method (stable)

Example: The Laplace operator

[edit]

The (continuous)Laplace operator inn{\displaystyle n}-dimensions is given byΔu(x)=i=1ni2u(x){\displaystyle \Delta u(x)=\sum _{i=1}^{n}\partial _{i}^{2}u(x)}.The discrete Laplace operatorΔhu{\displaystyle \Delta _{h}u} depends on the dimensionn{\displaystyle n}.

In 1D the Laplace operator is approximated asΔu(x)=u(x)u(xh)2u(x)+u(x+h)h2=:Δhu(x).{\displaystyle \Delta u(x)=u''(x)\approx {\frac {u(x-h)-2u(x)+u(x+h)}{h^{2}}}=:\Delta _{h}u(x)\,.}This approximation is usually expressed via the followingstencilΔh=1h2[121]{\displaystyle \Delta _{h}={\frac {1}{h^{2}}}{\begin{bmatrix}1&-2&1\end{bmatrix}}}and which represents a symmetric, tridiagonal matrix. For an equidistant grid one gets aToeplitz matrix.

The 2D case shows all the characteristics of the more general n-dimensional case. Each second partial derivative needs to be approximated similar to the 1D caseΔu(x,y)=uxx(x,y)+uyy(x,y)u(xh,y)2u(x,y)+u(x+h,y)h2+u(x,yh)2u(x,y)+u(x,y+h)h2=u(xh,y)+u(x+h,y)4u(x,y)+u(x,yh)+u(x,y+h)h2=:Δhu(x,y),{\displaystyle {\begin{aligned}\Delta u(x,y)&=u_{xx}(x,y)+u_{yy}(x,y)\\&\approx {\frac {u(x-h,y)-2u(x,y)+u(x+h,y)}{h^{2}}}+{\frac {u(x,y-h)-2u(x,y)+u(x,y+h)}{h^{2}}}\\&={\frac {u(x-h,y)+u(x+h,y)-4u(x,y)+u(x,y-h)+u(x,y+h)}{h^{2}}}\\&=:\Delta _{h}u(x,y)\,,\end{aligned}}}which is usually given by the followingstencilΔh=1h2[11411].{\displaystyle \Delta _{h}={\frac {1}{h^{2}}}{\begin{bmatrix}&1\\1&-4&1\\&1\end{bmatrix}}\,.}

Consistency

[edit]

Consistency of the above-mentioned approximation can be shown for highly regular functions, such asuC4(Ω){\displaystyle u\in C^{4}(\Omega )}.The statement isΔuΔhu=O(h2).{\displaystyle \Delta u-\Delta _{h}u={\mathcal {O}}(h^{2})\,.}

To prove this, one needs to substituteTaylor Series expansions up to order 3 into the discrete Laplace operator.

Properties

[edit]

Subharmonic

[edit]

Similar tocontinuous subharmonic functions one can definesubharmonic functions for finite-difference approximationsuh{\displaystyle u_{h}}Δhuh0.{\displaystyle -\Delta _{h}u_{h}\leq 0\,.}

Mean value

[edit]

One can define a generalstencil ofpositive type via[αNαWαCαEαS],αi>0,αC=i{N,E,S,W}αi.{\displaystyle {\begin{bmatrix}&\alpha _{N}\\\alpha _{W}&-\alpha _{C}&\alpha _{E}\\&\alpha _{S}\end{bmatrix}}\,,\quad \alpha _{i}>0\,,\quad \alpha _{C}=\sum _{i\in \{N,E,S,W\}}\alpha _{i}\,.}

Ifuh{\displaystyle u_{h}} is (discrete) subharmonic then the following mean value property holdsuh(xC)i{N,E,S,W}αiuh(xi)i{N,E,S,W}αi,{\displaystyle u_{h}(x_{C})\leq {\frac {\sum _{i\in \{N,E,S,W\}}\alpha _{i}u_{h}(x_{i})}{\sum _{i\in \{N,E,S,W\}}\alpha _{i}}}\,,}where the approximation is evaluated on points of the grid, and the stencil is assumed to be of positive type.

A similarmean value property also holds for the continuous case.

Maximum principle

[edit]

For a (discrete) subharmonic functionuh{\displaystyle u_{h}} the following holdsmaxΩhuhmaxΩhuh,{\displaystyle \max _{\Omega _{h}}u_{h}\leq \max _{\partial \Omega _{h}}u_{h}\,,}whereΩh,Ωh{\displaystyle \Omega _{h},\partial \Omega _{h}} are discretizations of the continuous domainΩ{\displaystyle \Omega }, respectively the boundaryΩ{\displaystyle \partial \Omega }.

A similarmaximum principle also holds for the continuous case.

The SBP-SAT method

[edit]

The SBP-SAT (summation by parts - simultaneous approximation term) method is a stable and accurate technique for discretizing and imposing boundary conditions of a well-posed linearpartial differential equation using high order finite differences.[8][9]

The method is based on finite differences where the differentiation operators exhibitsummation-by-parts properties. Typically, these operators consist of differentiation matrices with central difference stencils in the interior with carefully chosen one-sided boundary stencils designed to mimic integration-by-parts in the discrete setting. Using the SAT technique, the boundary conditions of the PDE are imposed weakly, where the boundary values are "pulled" towards the desired conditions rather than exactly fulfilled. If the tuning parameters (inherent to the SAT technique) are chosen properly, the resulting system of ODE's will exhibit similar energy behavior as the continuous PDE, i.e. the system has no non-physical energy growth. This guarantees stability if an integration scheme with a stability region that includes parts of the imaginary axis, such as the fourth orderRunge-Kutta method, is used. This makes the SAT technique an attractive method of imposing boundary conditions for higher order finite difference methods, in contrast to for example the injection method, which typically will not be stable if high order differentiation operators are used.

See also

[edit]

References

[edit]
  1. ^abChristian Grossmann; Hans-G. Roos; Martin Stynes (2007).Numerical Treatment of Partial Differential Equations. Springer Science & Business Media. p. 23.ISBN 978-3-540-71584-9.
  2. ^Arieh Iserles (2008).A first course in the numerical analysis of differential equations. Cambridge University Press. p. 23.ISBN 9780521734905.
  3. ^abHoffman JD; Frankel S (2001).Numerical methods for engineers and scientists. CRC Press, Boca Raton.
  4. ^abJaluria Y; Atluri S (1994). "Computational heat transfer".Computational Mechanics.14 (5):385–386.Bibcode:1994CompM..14..385J.doi:10.1007/BF00377593.S2CID 119502676.
  5. ^Majumdar P (2005).Computational methods for heat and mass transfer (1st ed.). Taylor and Francis, New York.
  6. ^Smith GD (1985).Numerical solution of partial differential equations: finite difference methods (3rd ed.). Oxford University Press.
  7. ^Crank, J.The Mathematics of Diffusion. 2nd Edition, Oxford, 1975, p. 143.
  8. ^Bo Strand (1994). "Summation by Parts for Finite Difference Approximations for d/dx".Journal of Computational Physics.110 (1):47–67.Bibcode:1994JCoPh.110...47S.doi:10.1006/jcph.1994.1005.
  9. ^Mark H. Carpenter; David I. Gottlieb; Saul S. Abarbanel (1994). "Time-stable boundary conditions for finite-difference schemes solving hyperbolic systems: Methodology and application to high-order compact schemes".Journal of Computational Physics.111 (2):220–236.Bibcode:1994JCoPh.111..220C.doi:10.1006/jcph.1994.1057.hdl:2060/19930013937.

Further reading

[edit]
Finite difference
Parabolic
Hyperbolic
Others
Finite volume
Finite element
Meshless/Meshfree
Domain decomposition
Others
Related
International
National
Other
Retrieved from "https://en.wikipedia.org/w/index.php?title=Finite_difference_method&oldid=1317617778"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp