23  Power Series

Highlights of this Chapter: we prove to marvelous results about power series: we show that they are differentiable (and get a formula for their derivative), and we also prove a formula about how to approximate functions well with a power series, and in the limit get a power series representation of a known function, in terms of its derivatives at a single point.

23.1 Differentiating Term-By-Term

The goal of this section is to prove that power series are differentiable, and that we can differentiate them term by term. That is, we seek to prove

(k0akxk)=k0(akxk)=k1kakxk1

Because a derivative is defined as a limit, this process of bringing the derivative inside the sum is really an exchange of limits: and we know the tool for that,Dominated Convergence! This applies quite generally so we give a general formulation and then apply it to power series

Theorem 23.1 Consider a infinite sum kfk(x) of functions which (1) converges on a domain D (2) has each fk differentiable on D. If there is a sequence Mk such that

  • Mk with |fk(x)|<MK, for all xD
  • The sum Mk is convergent.

Then, the sum kfk(x) is convergent, and (kfk(x))=kfk(x)

Proof. Recall the limit definition of the derivative (): (kfk(x))=limyxkfk(y)kfk(x)yx Writing each sum as the limit of finite sums, we may use the limit theorems (,) to combine this into a single sum limyxlimNk=0Nfk(y)limNk=0Nfk(x)yx=limyxlimNk=0Nfk(y)fk(x)yx

And now, rewriting the limit of partial sums as an infinite sum, we see (kfk(x))=limyxkfk(y)fk(x)yx

If we are justified in switching the limit and the sum via Dominated Convergence, this becomes

klimyxfk(y)fk(x)yx=kfk(x)

which is exactly what we want. Thus, all we need to do is justify that the conditions of Dominated Convergence are satisfied, for the terms appearing here. To be precise, this is a limit of functions, and we evaluate these by showing they exist for arbitrary sequences ynx with ynx. Choosing such a sequence and plugging in, we see we are really considering the sequence of terms limkfk(yn)fk(x)ynx.

Dominated convergence tells us we need to bound these terms fk(yn)fk(x)ynx above by some Mk. As part of the theorem hypothesis, we are given that there exists an Mk bounding the derivative of fk on D, so we just need to show that these suffice. For any xyn then fk(yn)fk(x)ynx measures the slope of the secant line of fk between x and yn, so by the Mean Value Theroem () there is some cn between x and yn with |fk(yn)fk(x)ynx|=|fk(cn)| Since |fk(c)|Mk by assumption (as cD), Mk is a bound for this difference quotient, as required.

Now recall our other assumption on the Mk: that Mk converges! This means we can apply dominated convergence bringing the limit inside, and

limnkfk(yn)fk(x)ynx=klimnfk(yn)fk(x)ynx

Since fk is differentiable and ynx, by definition this limit converges to the derivative fk(x). Thus, the limit of our sums is actually equal to kfk(x). And, as ynx was arbitrary, this holds for all such sequences. This means the limit defininig the derivative exists, and putting it all together, that (kfk(x))=kfk(x) as required.

Now we look to apply this to the specific case of power series kakxk within their intervals of convergence. The proof is much the same spirit as for continuity (where we also used dominated convergence), where we provide bounds Mk by looking at a larger point that remains in the interval of convergence. To do so, we need to understand the convergence of the power series of termwise derivatives. This is very similar to a previous homework problem (where you considered termwise antiderivatives) so we again leave as an exercise:

Exercise 23.1 Assume that kakxk has radius of convergence R. Show that kkakxk1 has the same radius of convergence.

Using this, we can put everything together.

Theorem 23.2 Let f=k0akxk be a power series with radius of convergence R. Then for x(R,R): f(x)=k1kakxk1

Proof. Let x(R,R) be arbitrary. Since x lies strictly within the interval of convergence, we may choose some closed interval (R,R) containing x. For concreteness we take [y,y] for some y<R, which we will use for the domain D when applying .

Getting to work verifying the assumptions of this theorem, our series converges on D (as D is a subset of the radius of convergence). Our individual functions fk of the sum are just the kth term of the series fk(x)=akxk. These are differentiable (as they are constant multiples of the monomial xk) with derivatives fk=kakxk1. The bounds Mk we seek are numbers which are greater in absolute value than this derivative on the domain D=[y,y]. Choose some value z>y within the radius of convergence (say, a=(R+y)/2). Then for all xD we have |x|<z and so

|x|k1<zk1=|zk1||kakxk=1|<|kakzk1|

So we may take Mk=|kakzk1|. But since the series of termwise derivatives has the same radius of convergence as the original series, and z is within the radius of convergence, we know kakzk1 converges absolutely! That is, Mk converges, and we are done.

Example 23.1 We know the geometric series converges to 1/(1x) on (1,1): k0xk=11x Differentiating term by term yields a power series for 1/(1x)2: 1(1x)2=(11x)=(k0xk)=k0xk=k1kxk1=1+2x+3x2+4x3+

The fact that power series are differentiable on their entire radius of convergence puts a strong constraint on which sort of functions can ever be written as the limit of such a series.

Example 23.2 The absolute value |x| is not expressible as a power series.

But this applies much more powerfully than even this: we can show that a power series must be infinitely differentiable at each point of its domain!

Corollary 23.1 (Power Series are Smooth) Proceed by induction on N. We know if kakxk is a power series it is at least N=1 times differentiable on the same radius of convergence, by our big result above. Now assume it is N times differentiable. Because we can differentiate power series term by term, the Nth derivative is also a power series, which has the same radius of convergence as the original.

But now we can apply our main theorem again: this power series is differentiable, with the same radius of convergence! Thus our original function is N+1 times differentiable, completing the induction.

23.2 Power Series Representations

While power series are interesting in their own right, our main purpose for them is to compute functions we already care about. In this section we use their differentiability to provide tools to do so.

Definition 23.1 (Power Series Representation) A power series representation of a function f at a point a is a power series p where p(x)=f(x) on some neighborhood of a.

Proposition 23.1 (Candidate Series Representation) Let f be a smooth real valued function whose domain contains a neighborhood of 0, and let p(x)=k0akxk be a power series which equals f on some neighborhood of zero. Then, the power series p is uniquely determined:

p(x)=k0f(k)(0)k!xk

Proof. Let f(x) be a smooth function and p(x)=k0akxk be a power series which equals f on some neighborhood of zero. Then in particular, p(0)=f(0), so

f(0)=limN(a0+a1x+a2x2++aNxN)=limN(a0+0+0++0)=a0

Now, we know the first coefficient of p. How can we get the next? Differentiate!

p(x)=(k0akxk)=k0(akxk)=k1kakxk1

Since f(x)=p(x) on some small neighborhood of zero and the derivative is a limit, f(0)=p(0). Evaluating this at 0 will give the constant term of the power series p

f(0)=limN(a1+2a2x+3a3x2+NaNxN1)=limN(a1+0+0++0)=a1

Continuing in this way, the second derivative will have a multiple of a2 as its constant term:

p(x)=2a2+32a3x+43a4x2+

And evaluating the equality f(x)=p(x) at zero yields

f(0)=2a2,soa2=f(0)2

This pattern continues indefinitely, as f is infinitely differentiable. The term an arrives in the constant term after n differentiations (as it was originally the coefficient of xn), at which point it becomes

anxnnanxn1n(n1)anxn2n(n1)(n2)321an

As the constant term of p(n) this means p(n)(0)=n!an, and so using f(n)(0)=p(n)(0), an=f(n)(0)n!

In each case there was no choice to be made, so long as f=p in any small neighborhood of zero, the unique formula for p is

p(x)=k0f(k)(0)k!xk

This candidate series makes it very easy to search for power series representations of known smooth functions: there’s only one series to even consider! This series is usually named after Brook Taylor, who gave their general formula in 1715.

Definition 23.2 (Taylor Series) For any smooth function f(x) we define the Taylor Polynomial (centered at 0) of degree N to be pN(x)=0kNf(k)(0)k!xk

In the limit as N, this defines the Taylor Series p(x) for f.

We’ve seen for example, that the geometric series k0xk is a power series representation of the function 1/(1x) at zero: it actually converges on the entire interval (1,1). There are many reasons one may be interested in finding a power series representation of a function - and the above theorem tells us that if we were to search for one, there is a single natural candidate. If there is any power series representation, its this one!

So the next natural step is to study this representation: does it actually converge to f(x)?

23.2.1 Taylor’s Error Formula

To prove that our series actually do what we want, we are going to need some tools relating a functions derivatives to its values. Rolle’s Theorem / the Mean Value Theorem does this for the first derivative, and so we present a generalization here the polynomial mean value theorem, which does so for nth derivatives.

Proposition 23.2 (A Generalized Rolle’s Theorem) Let f be a function which is n+1 times differentiable on the interior of an interval [a,b]. Assume that f(a)=f(b)=0, and further that the first n derivatives at a are zero: f(a)=f(a)=f(a)==f(n)(a)=0 Then, there exists some c(a,b) where f(n+1)(c)=0.

Proof. Because f is continuous and differentiable, and f(a)=f(b), the original Rolle’s Theorem implies that there exists some c(a,b) where f(c1)=0. But now, we know that f(a)=f(c1)=0, so we can apply Rolle’s theorem to f on $[a,c_1] to get a point c2(a,c1) with f(c2)=0.

Continuing in this way, we get a c3(a,c2) with f(3)(c)=0, all the way up to to a cn(a,cn1) where fn(cn)=0. This leaves one more application of Rolle’s theorem possible, as we assumed f(n)(a)=0, so we get a c(a,cn) with f(n+1)(c)=0 as claimed.

Proposition 23.3 (A Polynomial Mean Value Theorem) Let f(x) be an n+1-times differentiable function on [a,b] and h(x) a polynomial which shares the first n derivatives with f at zero: f(a)=h(a),f(a)=h(a),,f(n)(a)=p(n)(a) Then, if additionally f(b)=h(b), there must exist some point c(a,b) where f(n+1)(c)=h(n+1)(c)

Proof. Define the function g(x)=f(x)h(x). Then all the first n derivatives of g at x=a are zero (as f and h had the same derivatives), and furthermore g(b)=0 as well, since f(b)=h(b). This means we can apply the generalized Rolle’s theorem and find a c(a,b) with g(n+1)(c)=0 That is, f(n+1)(c)=h(n+1)(c).

Theorem 23.3 (Taylor Remainder) Let f(x) be an n+1-times differentiable function, and pn(x) the degree n Taylor polynomial p(x)=0knf(k)(0)k!xk.

Then for any fixed bR, we have f(b)=pn(b)+f(n+1)(c)(n+1)!bn+1

For some c[0,b].

Proof. Fix a point b, and consider the functions f(x) and pn(x) on the interval [0,b]. These share their first n derivatives at a, but f(b)pn(b): in fact, it is precisely this error we are trying to quantify.

We need to modify pn in some way without affecting its first n derivatives at zero. One natural way is to add a multiple of xn+1, so define

q(x)=pn(x)+λxn+1 for some λR, where we choose λ so that f(b)=q(b). Because we ensured q(k)(0)=f(k)(0) for kn, we can now apply the polynomial mean value theorem to these two functions, and get some c(0,b) where f(n+1)(c)=q(n+1)(c)

Since pn is degree n its n+1st derivative is zero, and q(n+1)(x)=0+(λxn+1)(n+1)=(n+1)!λ Putting these last two observations together yields

f(n+1)(c)=(n+1)!λλ=f(n+1)(c)(n+1)!

As q(b)=f(b) by construction, this in turn gives what we were after:

f(b)=pn(b)+f(n+1)(c)(n+1)!bn+1

23.2.2 Series Centered at aR

All of our discussion (and indeed, everything we will need about power series for our course) dealt with defining a power series based on derivative information at zero. But of course, this was an arbitrary choice: one could do exactly the same thing based at any point aR.

Theorem 23.4 Let f be a smooth function, defined in a neighborhood of aR. Then there is a unique power series which has all the same derivatives as f at a: p(x)=k0f(k)(a)k!(xa)k And, for any N the error between f and the Nth partial sum is quantified as f(x)pN(x)=f(N+1)(ξ)(N+1)!(xa)N+1 For some c[a,x].

Exercise 23.2 Prove this.