Movatterモバイル変換


[0]ホーム

URL:


Math Insight

To create your own interactive content like this, check out our new web sitedoenet.org!

Introduction to Taylor's theorem for multivariable functions

 

Remember one-variable calculus Taylor's theorem. Given a one variablefunction $f(x)$, you can fit it with a polynomial around $x=a$.

For example, the best linear approximation for $f(x)$ is\begin{align*} f(x) \approx f(a) + f\,'(a)(x-a).\end{align*}This linear approximation fits $f(x)$ (shown in green below) with aline (shown in blue) through $x=a$ thatmatches the slope of $f$ at $a$.

A tangent line to a graph

We can add additional, higher-order terms, to approximate $f(x)$ betternear $a$. The best quadratic approximation is\begin{align*} f(x) \approx f(a) + f\,'(a)(x-a) + \frac{1}{2} f\,''(a)(x-a)^2\end{align*}We could add third-order or even higher-order terms:\begin{align*} f(x) \approx f(a) + f\,'(a)(x-a) + \frac{1}{2} f\,''(a)(x-a)^2 + \frac{1}{6} f\,'''(a)(x-a)^3 + \cdots.\end{align*}The important point is that thisTaylor polynomialapproximates $f(x)$ well for $x$ near $a$.

We want to generalize the Taylor polynomial to (scalar-valued)functions of multiple variables:\begin{align*} f(\vc{x})= f(x_1,x_2, \ldots, x_n).\end{align*}

We already know the bestlinear approximation to $f$. It involves the derivative,\begin{align*} f(\vc{x}) \approx f(\vc{a}) + Df(\vc{a}) (\vc{x}-\vc{a}). \label{eq:firstorder}\end{align*}where $Df(\vc{a})$ is thematrix of partial derivatives. The linear approximation is the first-order Taylor polynomial.

What about the second-order Taylor polynomial? To find a quadraticapproximation, we need to add quadratic terms to our linearapproximation. For a function of one-variable $f(x)$, the quadraticterm was\begin{align*} \frac{1}{2} f\,''(a)(x-a)^2.\end{align*}For a function of multiple variables $f(\vc{x})$, what is analogous tothe second derivative?

Since $f(\vc{x})$ is scalar, the first derivative is $Df(\vc{x})$, a $1 \times n$ matrix,which we can view as an $n$-dimensional vector-valued function of the $n$-dimensional vector $\vc{x}$. For thesecond derivative of $f(\vc{x})$, we can take the matrix of partialderivatives of the function $Df(\vc{x})$. We could writeit as $DDf(\vc{x})$ for the moment. This secondderivative matrix is an $n \times n$ matrix called theHessian matrix of $f$. We'll denote it by $Hf(\vc{x})$,\begin{align*} Hf(\vc{x}) = DDf(\vc{x}).\end{align*}

When $f$ is a function of multiple variables, the second derivativeterm in the Taylor series will use the Hessian $Hf(\vc{a})$. For thesingle-variable case, we could rewrite the quadratic expressionas\begin{align*} \frac{1}{2} (x-a)f\,''(a)(x-a).\end{align*}The analog of this expression for the multivariable case is\begin{align*} \frac{1}{2} (\vc{x}-\vc{a})^T Hf(\vc{a}) (\vc{x}-\vc{a}).\end{align*}

We can add the aboveexpression to our first-order Taylor polynomialto obtain the second-order Taylor polynomial for functions of multiplevariables:\begin{align*} f(\vc{x}) \approx f(\vc{a}) + Df(\vc{a}) (\vc{x}-\vc{a}) + \frac{1}{2} (\vc{x}-\vc{a})^T Hf(\vc{a}) (\vc{x}-\vc{a}).\end{align*}The second-order Taylor polynomial is a better approximation of$f(\vc{x})$ near $\vc{x}=\vc{a}$ than is the linear approximation(which is the same as the first-order Taylor polynomial). We'll beable to use it for things such as finding a local minimum or localmaximum of the function $f(\vc{x})$.

You can read some exampleshere.


[8]ページ先頭

©2009-2025 Movatter.jp