Calculus/Taylor series

Taylor Series Edit

Definition: Taylor series

A function   is said to be analytic if it can be represented by the an infinite power series


The Taylor expansion or Taylor series representation of a function, then, is


  sin(x) and Taylor approximations, polynomials of degree 1, 3, 5, 7, 9, 11 and 13.

Here,   is the factorial of   and   denotes the  th derivative of   at the point   . If this series converges for every   in the interval   and the sum is equal to   , then the function   is called analytic. To check whether the series converges towards  , one normally uses estimates for the remainder term of Taylor's theorem. A function is analytic if and only if a power series converges to the function; the coefficients in that power series are then necessarily the ones given in the above Taylor series formula.

If   , the series is also called a Maclaurin series.

The importance of such a power series representation is threefold. First, differentiation and integration of power series can be performed term by term and is hence particularly easy. Second, an analytic function can be uniquely extended to a holomorphic function defined on an open disk in the complex plane, which makes the whole machinery of complex analysis available. Third, the (truncated) series can be used to approximate values of the function near the point of expansion.

  The function   is not analytic: the Taylor series is 0, although the function is not.

Note that there are examples of infinitely often differentiable functions   whose Taylor series converge, but are not equal to   . For instance, for the function defined piecewise by saying that   , all the derivatives are 0 at   , so the Taylor series of   is 0, and its radius of convergence is infinite, even though the function most definitely is not 0. This particular pathology does not afflict complex-valued functions of a complex variable. Notice that   does not approach 0 as   approaches 0 along the imaginary axis.

Some functions cannot be written as Taylor series because they have a singularity; in these cases, one can often still achieve a series expansion if one allows also negative powers of the variable  ; see Laurent series. For example,   can be written as a Laurent series.

The Parker-Sockacki theorem is a recent advance in finding Taylor series which are solutions to differential equations. This theorem is an expansion on the Picard iteration.

Derivation Edit

Suppose we want to represent a function as an infinite power series, or in other words a polynomial with infinite terms of degree "infinity". Each of these terms are assumed to have unique coefficients, as do most finite-polynomials do. We can represent this as an infinite sum like so:


where   is the radius of convergence and   are coefficients. Next, with summation notation, we can efficiently represent this series as


which will become more useful later. As of now, we have no schematic for finding the coefficients other than finding each one in the series by hand. That method would not be particularly useful. Let us, then, try to find a pattern and a general solution for finding the coefficients. As of now, we have a simple method for finding the first coefficient. If we substitute   for   then we get


This gives us   . This is useful, but we still would like a general equation to find any coefficient in the series. We can try differentiating with respect to x the series to get


We can assume   and   are constant. This proves to be useful, because if we again substitute   for   we get


Noting that the first derivative has one constant term ( ) we can find the second derivative to find   . It is


If we again substitute   for   :


Note that  's initial exponent was 2, and  's initial exponent was 1. This is slightly more enlightening, however it is still slightly ambiguous as to what is happening. Going off the previous examples, if we differentiate again we get


If we substitute   we, again, that


By now, the pattern should be becoming clearer.   looks suspiciously like   . And indeed, it is! If we carry this out   times by finding the  th derivative, we find that the multiple of the coefficient is   . So for some   , for any integer   ,


Or, with some simple manipulation, more usefully,


where   and   and so on. With this, we can find any coefficient of the "infinite polynomial". Using the summation definition for our "polynomial" given earlier,


we can substitute for   to get


This is the definition of any Taylor series. But now that we have this series, how can we derive the definition for a given analytic function? We can do just as the definition specifies, and fill in all the necessary information. But we will also want to find a specific pattern, because sometimes we are left with a great many terms simplifying to 0.

First, we have to find   . Because we are now deriving our own Taylor Series, we can choose anything we want for   , but note that not all functions will work. It would be useful to use a function that we can easily find the  -th derivative for. A good example of this would be   . With   chosen, we can begin to find the derivatives. Before we begin, we should also note that   is essentially the "offset" of the function along the x-axis, because this is also essentially true for any polynomial. With that in mind, we can assume, in this particular case, that the offset is   and so  . With that in mind, "0-th" derivative or the function itself would be


If we plug that in to the definition of the first term in the series, again noting that   , we get


where   . This means that the first term of the series is 0, because anything multiplied by 0 is 0. Take note that not all Taylor series start out with a 0 term. Next, to find the next term, we need to find the first derivative of the function. Remembering that the derivative of   is   we get that


This means that our second term in the series is


Next, we need to find the third term. We repeat this process.


Because the derivative of   . We continue with


The fourth term:


Repeating this process we can get the sequence


which simplifies to


Because we are ultimately dealing with a series, the zero terms can be ignored, giving use the new sequence


There is a pattern here, however it may be easier to see if we take the numerator and the denominator separately. The numerator:


And for the   part of the terms, we have the sequence


By this point, at least for the denominator and the   part, the pattern should be obvious. It is, for the denominator


The   term:


Finally, the numerator may not be as obvious, but it follows this pattern:


With all of these things discovered, we can put them together to find the rule for the  th term of the sequence:


And so our Taylor (Maclaurin) series for   is


List of Taylor series Edit

Several important Taylor series expansions follow. All these expansions are also valid for complex arguments   .

Exponential function and natural logarithm:


Geometric series:


Binomial series:


Trigonometric functions:


Hyperbolic functions:


Lambert's W function:


The numbers   appearing in the expansions of   and   are the Bernoulli numbers. The   in the binomial expansion are the binomial coefficients. The   in the expansion of   are Euler numbers.

Multiple dimensions Edit

The Taylor series may be generalized to functions of more than one variable with


History Edit

The Taylor series is named for mathematician Brook Taylor, who first published the power series formula in 1715.

Constructing a Taylor Series Edit

Several methods exist for the calculation of Taylor series of a large number of functions. One can attempt to use the Taylor series as-is and generalize the form of the coefficients, or one can use manipulations such as substitution, multiplication or division, addition or subtraction of standard Taylor series (such as those above) to construct the Taylor series of a function, by virtue of Taylor series being power series. In some cases, one can also derive the Taylor series by repeatedly applying integration by parts. The use of computer algebra systems to calculate Taylor series is common, since it eliminates tedious substitution and manipulation.

Example 1 Edit

Consider the function


for which we want a Taylor series at 0.

We have for the natural logarithm


and for the cosine function


We can simply substitute the second series into the first. Doing so gives


Expanding by using multinomial coefficients gives the required Taylor series. Note that cosine and therefore   are even functions, meaning that  , hence the coefficients of the odd powers  ,  ,  ,   and so on have to be zero and don't need to be calculated. The first few terms of the series are


The general coefficient can be represented using Faà di Bruno's formula. However, this representation does not seem to be particularly illuminating and is therefore omitted here.

Example 2 Edit

Suppose we want the Taylor series at 0 of the function


We have for the exponential function


and, as in the first example,


Assume the power series is


Then multiplication with the denominator and substitution of the series of the cosine yields


Collecting the terms up to fourth order yields


Comparing coefficients with the above series of the exponential function yields the desired Taylor series


Convergence Edit

Generalized Mean Value Theorem Edit