Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Differential of a function

From Wikipedia, the free encyclopedia
Notion in calculus
For other uses of "differential" in mathematics, seeDifferential (mathematics).
Part of a series of articles about
Calculus
abf(t)dt=f(b)f(a){\displaystyle \int _{a}^{b}f'(t)\,dt=f(b)-f(a)}

Incalculus, thedifferential represents theprincipal part of the change in afunctiony=f(x){\displaystyle y=f(x)} with respect to changes in the independent variable. The differentialdy{\displaystyle dy} is defined bydy=f(x)dx,{\displaystyle dy=f'(x)\,dx,}wheref(x){\displaystyle f'(x)} is thederivative off with respect tox{\displaystyle x}, anddx{\displaystyle dx} is an additional realvariable (so thatdy{\displaystyle dy} is a function ofx{\displaystyle x} anddx{\displaystyle dx}). The notation is such that the equation

dy=dydxdx{\displaystyle dy={\frac {dy}{dx}}\,dx}

holds, where the derivative is represented in theLeibniz notationdy/dx{\displaystyle dy/dx}, and this is consistent with regarding the derivative as the quotient of the differentials. One also writes

df(x)=f(x)dx.{\displaystyle df(x)=f'(x)\,dx.}

The precise meaning of the variablesdy{\displaystyle dy} anddx{\displaystyle dx} depends on the context of the application and the required level of mathematical rigor. The domain of these variables may take on a particular geometrical significance if the differential is regarded as a particulardifferential form, or analytical significance if the differential is regarded as alinear approximation to the increment of a function. Traditionally, the variablesdx{\displaystyle dx} anddy{\displaystyle dy} are considered to be very small (infinitesimal), and this interpretation is made rigorous innon-standard analysis.

History and usage

[edit]

The differential was first introduced via an intuitive or heuristic definition byIsaac Newton and furthered byGottfried Leibniz, who thought of the differential dy as an infinitely small (orinfinitesimal) change in the value y of the function, corresponding to an infinitely small change dx in the function's argument x. For that reason, the instantaneous rate of change ofy with respect tox, which is the value of thederivative of the function, is denoted by the fraction

dydx{\displaystyle {\frac {dy}{dx}}}in what is called theLeibniz notation for derivatives. The quotientdy/dx{\displaystyle dy/dx} is not infinitely small; rather it is areal number.

The use of infinitesimals in this form was widely criticized, for instance by the famous pamphletThe Analyst by Bishop Berkeley.Augustin-Louis Cauchy (1823) defined the differential without appeal to the atomism of Leibniz's infinitesimals.[1][2] Instead, Cauchy, followingd'Alembert, inverted the logical order of Leibniz and his successors: the derivative itself became the fundamental object, defined as alimit of difference quotients, and the differentials were then defined in terms of it. That is, one was free todefine the differentialdy{\displaystyle dy} by an expressiondy=f(x)dx{\displaystyle dy=f'(x)\,dx}in whichdy{\displaystyle dy} anddx{\displaystyle dx} are simply new variables taking finite real values,[3] not fixed infinitesimals as they had been for Leibniz.[4]

According toBoyer (1959, p. 12), Cauchy's approach was a significant logical improvement over the infinitesimal approach of Leibniz because, instead of invoking the metaphysical notion of infinitesimals, the quantitiesdy{\displaystyle dy} anddx{\displaystyle dx} could now be manipulated in exactly the same manner as any other real quantitiesin a meaningful way. Cauchy's overall conceptual approach to differentials remains the standard one in modern analytical treatments,[5] although the final word on rigor, a fully modern notion of the limit, was ultimately due toKarl Weierstrass.[6]

In physical treatments, such as those applied to the theory ofthermodynamics, the infinitesimal view still prevails.Courant & John (1999, p. 184) reconcile the physical use of infinitesimal differentials with the mathematical impossibility of them as follows. The differentials represent finite non-zero values that are smaller than the degree of accuracy required for the particular purpose for which they are intended. Thus "physical infinitesimals" need not appeal to a corresponding mathematical infinitesimal in order to have a precise sense.

Following twentieth-century developments inmathematical analysis anddifferential geometry, it became clear that the notion of the differential of a function could be extended in a variety of ways. Inreal analysis, it is more desirable to deal directly with the differential as the principal part of the increment of a function. This leads directly to the notion that the differential of a function at a point is alinear functional of an incrementΔx{\displaystyle \Delta x}. This approach allows the differential (as a linear map) to be developed for a variety of more sophisticated spaces, ultimately giving rise to such notions as theFréchet orGateaux derivative. Likewise, indifferential geometry, the differential of a function at a point is a linear function of atangent vector (an "infinitely small displacement"), which exhibits it as a kind of one-form: theexterior derivative of the function. Innon-standard calculus, differentials are regarded as infinitesimals, which can themselves be put on a rigorous footing (seedifferential (infinitesimal)).

Definition

[edit]
The differential of a functionf(x){\displaystyle f(x)} at a pointx0{\displaystyle x_{0}}.

The differential is defined in modern treatments of differential calculus as follows.[7] The differential of a functionf(x){\displaystyle f(x)} of a single real variablex{\displaystyle x} is the functiondf{\displaystyle df} of two independent real variablesx{\displaystyle x} andΔx{\displaystyle \Delta x} given by

df(x,Δx) =def f(x)Δx.{\displaystyle df(x,\Delta x)\ {\stackrel {\mathrm {def} }{=}}\ f'(x)\,\Delta x.}

One or both of the arguments may be suppressed, i.e., one may seedf(x){\displaystyle df(x)} or simplydf{\displaystyle df}. Ify=f(x){\displaystyle y=f(x)}, the differential may also be written asdy{\displaystyle dy}. Sincedx(x,Δx)=Δx{\displaystyle dx(x,\Delta x)=\Delta x}, it is conventional to writedx=Δx{\displaystyle dx=\Delta x} so that the following equality holds:

df(x)=f(x)dx{\displaystyle df(x)=f'(x)\,dx}

This notion of differential is broadly applicable when alinear approximation to a function is sought, in which the value of the incrementΔx{\displaystyle \Delta x} is small enough. More precisely, iff{\displaystyle f} is adifferentiable function atx{\displaystyle x}, then the difference iny{\displaystyle y}-values

Δy =def f(x+Δx)f(x){\displaystyle \Delta y\ {\stackrel {\rm {def}}{=}}\ f(x+\Delta x)-f(x)}

satisfies

Δy=f(x)Δx+ε=df(x)+ε{\displaystyle \Delta y=f'(x)\,\Delta x+\varepsilon =df(x)+\varepsilon \,}

where the errorε{\displaystyle \varepsilon } in the approximation satisfiesε/Δx0{\displaystyle \varepsilon /\Delta x\rightarrow 0} asΔx0{\displaystyle \Delta x\rightarrow 0}. In other words, one has the approximate identity

Δydy{\displaystyle \Delta y\approx dy}

in which the error can be made as small as desired relative toΔx{\displaystyle \Delta x} by constrainingΔx{\displaystyle \Delta x} to be sufficiently small; that is to say,ΔydyΔx0{\displaystyle {\frac {\Delta y-dy}{\Delta x}}\to 0}asΔx0{\displaystyle \Delta x\rightarrow 0}. For this reason, the differential of a function is known as theprincipal (linear) part in the increment of a function: the differential is alinear function of the incrementΔx{\displaystyle \Delta x}, and although the errorε{\displaystyle \varepsilon } may be nonlinear, it tends to zero rapidly asΔx{\displaystyle \Delta x} tends to zero.

Differentials in several variables

[edit]
Operator / Functionf(x){\displaystyle f(x)}f(x,y,u(x,y),v(x,y)){\displaystyle f(x,y,u(x,y),v(x,y))}
Differential1:df=deffxdx{\displaystyle df\,{\overset {\underset {\mathrm {def} }{}}{=}}\,f'_{x}\,dx}2:dxf=deffxdx{\displaystyle d_{x}f\,{\overset {\underset {\mathrm {def} }{}}{=}}\,f'_{x}\,dx}

3:df=deffxdx+fydy+fudu+fvdv{\displaystyle df\,{\overset {\underset {\mathrm {def} }{}}{=}}\,f'_{x}dx+f'_{y}dy+f'_{u}du+f'_{v}dv}

Partial derivativefx=(1)dfdx{\displaystyle f'_{x}\,{\overset {\underset {\mathrm {(1)} }{}}{=}}\,{\frac {df}{dx}}}fx=(2)dxfdx=fx{\displaystyle f'_{x}\,{\overset {\underset {\mathrm {(2)} }{}}{=}}\,{\frac {d_{x}f}{dx}}={\frac {\partial f}{\partial x}}}
Total derivativedfdx=(1)fx{\displaystyle {\frac {df}{dx}}\,{\overset {\underset {\mathrm {(1)} }{}}{=}}\,f'_{x}}dfdx=(3)fx+fududx+fvdvdx;(fydydx=0){\displaystyle {\frac {df}{dx}}\,{\overset {\underset {\mathrm {(3)} }{}}{=}}\,f'_{x}+f'_{u}{\frac {du}{dx}}+f'_{v}{\frac {dv}{dx}};(f'_{y}{\frac {dy}{dx}}=0)}

FollowingGoursat (1904, I, §15), for functions of more than one independent variable,

y=f(x1,,xn),{\displaystyle y=f(x_{1},\dots ,x_{n}),}

thepartial differential ofy with respect to any one of the variables xi is the principal part of the change iny resulting from a change dxi in that one variable. The partial differential is therefore

yxidxi{\displaystyle {\frac {\partial y}{\partial x_{i}}}dx_{i}}

involving thepartial derivative ofy with respect to xi. The sum of the partial differentials with respect to all of the independent variables is thetotal differential

dy=yx1dx1++yxndxn,{\displaystyle dy={\frac {\partial y}{\partial x_{1}}}dx_{1}+\cdots +{\frac {\partial y}{\partial x_{n}}}dx_{n},}

which is the principal part of the change iny resulting from changes in the independent variables xi.

More precisely, in the context of multivariable calculus, followingCourant (1937b), iff is a differentiable function, then by thedefinition of differentiability, the increment

Δy =def f(x1+Δx1,,xn+Δxn)f(x1,,xn)=yx1Δx1++yxnΔxn+ε1Δx1++εnΔxn{\displaystyle {\begin{aligned}\Delta y&{}~{\stackrel {\mathrm {def} }{=}}~f(x_{1}+\Delta x_{1},\dots ,x_{n}+\Delta x_{n})-f(x_{1},\dots ,x_{n})\\&{}={\frac {\partial y}{\partial x_{1}}}\Delta x_{1}+\cdots +{\frac {\partial y}{\partial x_{n}}}\Delta x_{n}+\varepsilon _{1}\Delta x_{1}+\cdots +\varepsilon _{n}\Delta x_{n}\end{aligned}}}

where the error termsε i tend to zero as the incrementsΔxi jointly tend to zero. The total differential is then rigorously defined as

dy=yx1Δx1++yxnΔxn.{\displaystyle dy={\frac {\partial y}{\partial x_{1}}}\Delta x_{1}+\cdots +{\frac {\partial y}{\partial x_{n}}}\Delta x_{n}.}

Since, with this definition,dxi(Δx1,,Δxn)=Δxi,{\displaystyle dx_{i}(\Delta x_{1},\dots ,\Delta x_{n})=\Delta x_{i},}one hasdy=yx1dx1++yxndxn.{\displaystyle dy={\frac {\partial y}{\partial x_{1}}}\,dx_{1}+\cdots +{\frac {\partial y}{\partial x_{n}}}\,dx_{n}.}

As in the case of one variable, the approximate identity holds

dyΔy{\displaystyle dy\approx \Delta y}

in which the total error can be made as small as desired relative toΔx12++Δxn2{\textstyle {\sqrt {\Delta x_{1}^{2}+\cdots +\Delta x_{n}^{2}}}} by confining attention to sufficiently small increments.

Application of the total differential to error estimation

[edit]

In measurement, the total differential is used inestimating the errorΔf{\displaystyle \Delta f} of a functionf{\displaystyle f} based on the errorsΔx,Δy,{\displaystyle \Delta x,\Delta y,\ldots } of the parametersx,y,{\displaystyle x,y,\ldots }. Assuming that the interval is short enough for the change to be approximately linear:

Δf(x)=f(x)Δx{\displaystyle \Delta f(x)=f'(x)\Delta x}

and that all variables are independent, then for all variables,

Δf=fxΔx+fyΔy+{\displaystyle \Delta f=f_{x}\Delta x+f_{y}\Delta y+\cdots }

This is because the derivativefx{\displaystyle f_{x}} with respect to the particular parameterx{\displaystyle x} gives the sensitivity of the functionf{\displaystyle f} to a change inx{\displaystyle x}, in particular the errorΔx{\displaystyle \Delta x}. As they are assumed to be independent, the analysis describes the worst-case scenario. The absolute values of the component errors are used, because after simple computation, the derivative may have a negative sign. From this principle the error rules of summation, multiplication etc. are derived, e.g.:

Letf(a,b)=ab{\displaystyle f(a,b)=ab}. Then, the finite error can be approximated as

Δf=faΔa+fbΔb.{\displaystyle \Delta f=f_{a}\Delta a+f_{b}\Delta b.}Evaluating the derivatives:Δf=bΔa+aΔb.{\displaystyle \Delta f=b\Delta a+a\Delta b.}Dividing byf, which isa ×b

Δff=Δaa+Δbb{\displaystyle {\frac {\Delta f}{f}}={\frac {\Delta a}{a}}+{\frac {\Delta b}{b}}}

That is to say, in multiplication, the totalrelative error is the sum of the relative errors of the parameters.

To illustrate how this depends on the function considered, consider the case where the function isf(a,b)=alnb{\displaystyle f(a,b)=a\ln b} instead. Then, it can be computed that the error estimate isΔff=Δaa+Δbblnb{\displaystyle {\frac {\Delta f}{f}}={\frac {\Delta a}{a}}+{\frac {\Delta b}{b\ln b}}}with an extralnb factor not found in the case of a simple product. This additional factor tends to make the error smaller, as the denominatorb lnb is larger than a bare b.

Higher-order differentials

[edit]

Higher-order differentials of a functiony =f(x) of a single variablex can be defined via:[8]d2y=d(dy)=d(f(x)dx)=(df(x))dx=f(x)(dx)2,{\displaystyle d^{2}y=d(dy)=d(f'(x)dx)=(df'(x))dx=f''(x)\,(dx)^{2},}and, in general,dny=f(n)(x)(dx)n.{\displaystyle d^{n}y=f^{(n)}(x)\,(dx)^{n}.}Informally, this motivates Leibniz's notation for higher-order derivativesf(n)(x)=dnfdxn.{\displaystyle f^{(n)}(x)={\frac {d^{n}f}{dx^{n}}}.}When the independent variablex itself is permitted to depend on other variables, then the expression becomes more complicated, as it must include also higher order differentials inx itself. Thus, for instance,d2y=f(x)(dx)2+f(x)d2xd3y=f(x)(dx)3+3f(x)dxd2x+f(x)d3x{\displaystyle {\begin{aligned}d^{2}y&=f''(x)\,(dx)^{2}+f'(x)d^{2}x\\[1ex]d^{3}y&=f'''(x)\,(dx)^{3}+3f''(x)dx\,d^{2}x+f'(x)d^{3}x\end{aligned}}}and so forth.

Similar considerations apply to defining higher order differentials of functions of several variables. For example, iff is a function of two variablesx andy, thendnf=k=0n(nk)nfxkynk(dx)k(dy)nk,{\displaystyle d^{n}f=\sum _{k=0}^{n}{\binom {n}{k}}{\frac {\partial ^{n}f}{\partial x^{k}\partial y^{n-k}}}(dx)^{k}(dy)^{n-k},}where(nk){\textstyle {\binom {n}{k}}} is abinomial coefficient. In more variables, an analogous expression holds, but with an appropriatemultinomial expansion rather than binomial expansion.[9]

Higher order differentials in several variables also become more complicated when the independent variables are themselves allowed to depend on other variables. For instance, for a functionf ofx andy which are allowed to depend on auxiliary variables, one hasd2f=(2fx2(dx)2+22fxydxdy+2fy2(dy)2)+fxd2x+fyd2y.{\displaystyle d^{2}f=\left({\frac {\partial ^{2}f}{\partial x^{2}}}(dx)^{2}+2{\frac {\partial ^{2}f}{\partial x\partial y}}dx\,dy+{\frac {\partial ^{2}f}{\partial y^{2}}}(dy)^{2}\right)+{\frac {\partial f}{\partial x}}d^{2}x+{\frac {\partial f}{\partial y}}d^{2}y.}

Because of this notational awkwardness, the use of higher order differentials was roundly criticized byHadamard (1935), who concluded:

Enfin, que signifie ou que représente l'égalité

d2z=rdx2+2sdxdy+tdy2?{\displaystyle d^{2}z=r\,dx^{2}+2s\,dx\,dy+t\,dy^{2}\,?}

A mon avis, rien du tout.

That is:Finally, what is meant, or represented, by the equality [...]? In my opinion, nothing at all. In spite of this skepticism, higher order differentials did emerge as an important tool in analysis.[10]

In these contexts, then-th order differential of the functionf applied to an incrementΔx is defined bydnf(x,Δx)=dndtnf(x+tΔx)|t=0{\displaystyle d^{n}f(x,\Delta x)=\left.{\frac {d^{n}}{dt^{n}}}f(x+t\Delta x)\right|_{t=0}}or an equivalent expression, such aslimt0ΔtΔxnftn{\displaystyle \lim _{t\to 0}{\frac {\Delta _{t\Delta x}^{n}f}{t^{n}}}}whereΔtΔxnf{\displaystyle \Delta _{t\Delta x}^{n}f} is annthforward difference with incrementtΔx.

This definition makes sense as well iff is a function of several variables (for simplicity taken here as a vector argument). Then then-th differential defined in this way is ahomogeneous function of degreen in the vector incrementΔx. Furthermore, theTaylor series off at the pointx is given byf(x+Δx)f(x)+df(x,Δx)+12d2f(x,Δx)++1n!dnf(x,Δx)+{\displaystyle f(x+\Delta x)\sim f(x)+df(x,\Delta x)+{\frac {1}{2}}d^{2}f(x,\Delta x)+\cdots +{\frac {1}{n!}}d^{n}f(x,\Delta x)+\cdots }The higher orderGateaux derivative generalizes these considerations to infinite dimensional spaces.

Properties

[edit]

A number of properties of the differential follow in a straightforward manner from the corresponding properties of the derivative, partial derivative, and total derivative. These include:[11]

An operationd with these two properties is known inabstract algebra as aderivation. They imply the power ruled(fn)=nfn1df{\displaystyle d(f^{n})=nf^{n-1}df}In addition, various forms of thechain rule hold, in increasing level of generality:[12]

General formulation

[edit]
See also:Fréchet derivative andGateaux derivative

A consistent notion of differential can be developed for a functionf :RnRm between twoEuclidean spaces. LetxxRn be a pair ofEuclidean vectors. The increment in the functionf isΔf=f(x+Δx)f(x).{\displaystyle \Delta f=f(\mathbf {x} +\Delta \mathbf {x} )-f(\mathbf {x} ).}If there exists anm ×nmatrixA such thatΔf=AΔx+Δxε{\displaystyle \Delta f=A\Delta \mathbf {x} +\|\Delta \mathbf {x} \|{\boldsymbol {\varepsilon }}}in which the vectorε → 0 asΔx → 0, thenf is by definition differentiable at the pointx. The matrixA is sometimes known as theJacobian matrix, and thelinear transformation that associates to the incrementΔxRn the vectorAΔxRm is, in this general setting, known as the differentialdf(x) off at the pointx. This is precisely theFréchet derivative, and the same construction can be made to work for a function between anyBanach spaces.

Another fruitful point of view is to define the differential directly as a kind ofdirectional derivative:df(x,h)=limt0f(x+th)f(x)t=ddtf(x+th)|t=0,{\displaystyle df(\mathbf {x} ,\mathbf {h} )=\lim _{t\to 0}{\frac {f(\mathbf {x} +t\mathbf {h} )-f(\mathbf {x} )}{t}}=\left.{\frac {d}{dt}}f(\mathbf {x} +t\mathbf {h} )\right|_{t=0},}which is the approach already taken for defining higher order differentials (and is most nearly the definition set forth by Cauchy). Ift represents time andx position, thenh represents a velocity instead of a displacement as we have heretofore regarded it. This yields yet another refinement of the notion of differential: that it should be a linear function of a kinematic velocity. The set of all velocities through a given point of space is known as thetangent space, and sodf gives a linear function on the tangent space: adifferential form. With this interpretation, the differential off is known as theexterior derivative, and has broad application indifferential geometry because the notion of velocities and the tangent space makes sense on anydifferentiable manifold. If, in addition, the output value off also represents a position (in a Euclidean space), then a dimensional analysis confirms that the output value ofdf must be a velocity. If one treats the differential in this manner, then it is known as thepushforward since it "pushes" velocities from a source space into velocities in a target space.

Other approaches

[edit]
Main article:Differential (infinitesimal)

Although the notion of having an infinitesimal incrementdx is not well-defined in modernmathematical analysis, a variety of techniques exist for defining theinfinitesimal differential so that the differential of a function can be handled in a manner that does not clash with theLeibniz notation. These include:

Examples and applications

[edit]

Differentials may be effectively used innumerical analysis to study the propagation of experimental errors in a calculation, and thus the overallnumerical stability of a problem (Courant 1937a). Suppose that the variablex represents the outcome of an experiment andy is the result of a numerical computation applied tox. The question is to what extent errors in the measurement ofx influence the outcome of the computation ofy. If thex is known to within Δx of its true value, thenTaylor's theorem gives the following estimate on the error Δy in the computation ofy:Δy=f(x)Δx+(Δx)22f(ξ){\displaystyle \Delta y=f'(x)\Delta x+{\frac {(\Delta x)^{2}}{2}}f''(\xi )}whereξ =x +θΔx for some0 <θ < 1. IfΔx is small, then the second order term is negligible, so that Δy is, for practical purposes, well-approximated bydy =f'(x) Δx.

The differential is often useful to rewrite adifferential equationdydx=g(x){\displaystyle {\frac {dy}{dx}}=g(x)}in the formdy=g(x)dx,{\displaystyle dy=g(x)\,dx,}in particular when one wants toseparate the variables.

Notes

[edit]
  1. ^For a detailed historical account of the differential, seeBoyer 1959, especially page 275 for Cauchy's contribution on the subject. An abbreviated account appears inKline 1972, Chapter 40.
  2. ^Cauchy explicitly denied the possibility of actual infinitesimal and infinite quantities (Boyer 1959, pp. 273–275), and took the radically different point of view that "a variable quantity becomes infinitely small when its numerical value decreases indefinitely in such a way as to converge to zero" (Cauchy 1823, p. 12; translation fromBoyer 1959, p. 273).
  3. ^Boyer 1959, p. 275
  4. ^Boyer 1959, p. 12: "The differentials as thus defined are only newvariables, and not fixed infinitesimals..."
  5. ^Courant 1937a, II, §9: "Here we remark merely in passing that it is possible to use this approximate representation of the incrementΔy{\displaystyle \Delta y} by the linear expressionhf(x){\displaystyle hf(x)} to construct a logically satisfactory definition of a "differential", as was done by Cauchy in particular."
  6. ^Boyer 1959, p. 284
  7. ^See, for instance, the influential treatises ofCourant 1937a,Kline 1977,Goursat 1904, andHardy 1908. Tertiary sources for this definition include alsoTolstov 2001 andItô 1993, §106.
  8. ^Cauchy 1823. See also, for instance,Goursat 1904, I, §14.
  9. ^Goursat 1904, I, §14
  10. ^In particular toinfinite dimensional holomorphy (Hille & Phillips 1974) andnumerical analysis via the calculus offinite differences.
  11. ^Goursat 1904, I, §17
  12. ^Goursat 1904, I, §§14,16
  13. ^Eisenbud & Harris 1998.
  14. ^SeeKock 2006 andMoerdijk & Reyes 1991.
  15. ^SeeRobinson 1996 andKeisler 1986.

See also

[edit]

References

[edit]

External links

[edit]
Retrieved from "https://en.wikipedia.org/w/index.php?title=Differential_of_a_function&oldid=1293145252"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp