Exponentials
Proposition 24.1 Let be an exponential function. Then is differentiable on the entire real line, and
First we show that this formula holds so long as is actually differentiable at zero. Thus, differentiability at a single point is enough to ensure differentiability everywhere and fully determine the formula!
Proof. Let , and . Then we compute by the following limit:
Using the property of exponentials and the limit laws, we can factor an out of the entire numerator:
But, so the limit here is actually the *derivative of at zero$!
Next, we tackle the slightly more subtle problem of showing that is in fact differentiable at zero. This is tricky because all we have assumed is that is continuous and satisfies the law of exponents: how are we going to pull differentiability out of this? One trick is two parts (1) show the right and left hand limits defining the derivative exist, and (2) show they’re equal.
In fact, is a known number, its the natural log of (Cite where we’ll prove this later)
Proof. Because is convex (Exercise 20.5) so the difference quotient is monotone increasing and so the limit exists (as a sup) and exists (as an inf). Now that we know each of these limits exist, let’s show they are equal using the definition: To compute the lower limit, we can choose any sequence approaching from below: let be a positive sequence with , then will do:
As , we compute
But, since is continuous (by definition) and the limit theorems imply Thus, But this last limit evaluates exactly to the limit from above since and . Stringing all of this together, we finally see As both one sided limits exist and are equal the entire limit exists: is differentiable at .
When studying the functional equations for logs and exponentials we saw there is not one solution but a whole family of them. While the functional equation itself gave no preference to any exponential over any other, the derivative selects an obvious natural candidate:
Definition 24.1 We write for the exponential function which has . This exponential satisfies the simple differential identity
Note that by the chain rule we know such a thing exists so long as any exponential exists. If is any exponential then has derivative at !
A Series Representation
To work with the natural exponential efficiently, we need to find a formula that lets us compute it. And this is exactly what power series are good at! However, the theory of power series is a little tricky, as we saw in the last chapter. Not every function has a power series representation, but if a function does, there’s only one possibility:
Proposition 24.2 If the natural exponential has a power series representation, then it is
Proof. We know the only candidate series for a function is , so for this is
However, we know that and so inductively , and so Thus
Since is a power series, this really means that the limit of its partial sums equals , or
For any finite partial sum , we know that it is not exactly equal to (as this finite sum is just a polynomial!). Thus there must be some error term , or
This is helpful, as we know from the previous chapter how to calculate such an error, using the Taylor Error Formula: for each fixed and each fixed , there is some point such that
And, to show the power series becomes the natural exponential in the limit, we just need to show this error tends to zero!
Proposition 24.3 As , for any the Taylor error term for the exponential goes to zero:
Proof. Fix some . Then for an arbitrary , we know where is some number that we don’t have much control of (as it came from an existence proof: Rolle’s theorem in our derivation of the Taylor error). Because we don’t know explicitly, its hard to directly compute the limit and so instead we use the squeeze theorem:
We know that is an increasing function: thus, the fact that implies that , and multiplying this inequality through by yields the inequality
(Here I have assumed that : if then the inequalities reverse for even values of as is negative and we are multiplying through by a negative number. But this does not affect the fact that the error term is still sandwiched between the two.)
So now our problem reduces to showing that the upper and lower bounds converge to zero. Since is a constant (remember, is our variable here as we take the limit), a limit of both the upper and lower bounds comes down to just finding the limit
But this is just the st term of the power series we studied above! And since this power series converges, we know that as its terms must go to zero (the divergence test). Thus
and so by the squeeze theorem, converges and
Now we have all the components together at last: we know that exists, we have a candidate power series representation, that candidate converges, and the error between it and the exponential goes to zero!
Theorem 24.1 The natural exponential is given by the following power series
Proof. Fix an arbitrary . Then for any we can write For the partial sum of and the error. Since we have proven both and converge, we can take the limit of both sides using the limit theorems (and, as is constant in , clearly ):
Its incredible in and of itself to have such a simple, explicit formula for the natural exponential. But this is just the beginning: this series actually gives us a means to express all exponentials:
Theorem 24.2 Let be an arbitrary exponential function. Then has a power series representation on all of which can be expressed for some real nonzero as
Proof. Because is an exponential we know is differentiable, and that for all . Note that is nonzero; else we would have constantly, and so would be constant. Set .
Now, inductively take derivatives at zero:
Thus, if has a power series representation it must be
This is just the series for evaluated at : since exists and is an exponential, so is this function (as its defined just by a substitution). So there is such an exponential.
The Number
Recalling our work with irrational exponents, we know that exponentials are powers: if is an exponential with , then we may write for any (defined as a limit of rational exponents). So, our special exponential comes with a special number as its base.
Definition 24.2 We denote by the letter the base of the exponential : that is, , and
What is this natural base? We can estimate its value using the power series representation for , and the Taylor error formula.
Proposition 24.4 The base of the natural exponential is between and .
Proof. The series defining is all positive terms, so we see that is greater than any partial sum. Thus so we have the lower bound. To get the upper bound, we need to come up wtih a computable upper bound for our series. This turns out to be not that difficult: as the factorial grows so quickly, we can produce many upper bounds by just fining something that grows slower than the reciprocal and summing up their reciprocals. For instance, when
and so,
But this upper bound now is our favorite telescoping series! After a rewrite with partial fractions, we directly see that it sums to . Plugging this in,
How can we get a better estimate? Since we do have a convergent infinite series just sitting here defining for us, the answer seems obvious - why don’t we just sum up more and more terms of the series? And of course - that is part of the correct strategy, but it’s missing one key piece. If you add up the first 10 terms of the series and you get some number, how can you know how accurate this is?
Just because the first two digits are , who is to say that after adding a million more terms (all of which are positive) it won’t eventually become ? To give us any confidence in the value of we need a way of measuring how far off any of our partial sums could be.
Our usual approach is to try and produce sequences of upper and lower estimates: nested intervals of error bars to help us out. But here we have only one sequence (and producing even a single upper bound above was a bit of work!) so we need to look elsewhere. It turns out, the correct tool for the job is the Taylor Error formula once more!
Proposition 24.5 Adding up the first terms of the series expansion of results in a an estimate of the true value accurate to within .
Proof. The number is defined as , and so using we are just looking at the old equation
Where for . Since is increasing, we can bound below by and above by , and above by : thus
And so, the difference is bounded above by the upper bound
This gives us a readily computable, explicit estimate. Precisely adding up to the th term of the series yields
with the total error between this and is less than . Thus we can be confident that the first digit after the decimal is a 7, as .
Adding up five more terms, to gives
now with a maximal error of . This means we are now absolutely confident in the first six digits:
Pretty good, for only having to add eleven fractions together! Thats the sort of calculation one could even manage by hand.
Logarithms
Having done all this work for exponentials, we can immediately confirm that logarithms are also differentiable, and give a simple formula.
Proposition 24.6 Let be a logarithm function, then is differentiable and
Proof. Let be a logarithm, with inverse the exponential . We know that is differentiable and as its a constant multiple of the everywhere-positive itself. Thus by Theorem 21.7 the function is also differentiable. Choosing and setting gives Where the last equality follows as implies
Now we apply the theorem on differentiability of inverses one more time to remove the mention of in the final answer, and express everything in terms of the logarithm itself. Since and , we have and substituting this in gives the claimed form.
Thus every logarithm has as its derivative some multiple of . While the functional equation doesn’t distinguish between these different logarithms, calculus finds one of them most natural: when this constant is equal to 1!
Definition 24.3 (The Natural Log) We write for the logarithm function which has . This logarithm satisfies the simple differential identity
Furthermore the two notions of “naturalness” picking out a logarithm and an exponential are compatible with one another!
Corollary 24.1 The natural exponential and natural log are inverses of one another.
We will make much use of this pair of special functions, and their exeedingly simple differentiation rules. As a first application, we give a re-proof of the power rule avoiding difficult limiting arguments:
Theorem 24.3 (The General Power Rule) If and . Then is differentiable for all , and
Proof. Let be the natural exponential, and be the natural log. Then , and so . Using the property of logarithms and powers (?cor-log-exponent) this simplifies
By the chain rule,
But, recalling that this simplifies to
Unfortunately our newfound tool does not apply so well to giving a formula for the logarithm: power series are always defined on some symmetric interval about , but the domain of the logarithm is . Thus there is no simple power series that will equal !. We will come up with formulas to compute the logarithm later on, first as an integral; and then as a series (that converges only for some values of ).
Trigonometric Functions
Recall our definition of trigonometric functions: two functions are trigonometric if they satisfy the angle difference identities
In one option for the final project in this class, you prove that such functions exist by taking a proposed power series, and showing directly that it satisfies these identities. Our goal in this section is to build up the work that leads to proposing that power series via a sequence of exercises
Exercise 24.1 Assume that a pair of trigonometric functions are differentiable at zero (so , exist). Use the functional equations defining them to show that they are differentiable at every .
Hint: Write out the difference quotients, as a limit for . Additionally, we know are continuous by definition, and have previously shown , .
We will not prove that and are differentiable at zero here (though that can be done, much like we did for the exponential above!). Since our goal is simply to propose a power series,
Exercise 24.2 Prove that if are differentiable at zero, then and .
Then deduce the following differentiation laws:
for some nonzero constant.
Hint: recall the definition that are nonconstant and continuous. Also recall Fermat’s Theorem.
This singles out an obvious ‘best’ set of trigonometric functions, setting the arbitrary constant parameter that appears equal to one!
Definition 24.4 (The Natural Trigonometric Functions) The natural trigonometric functions are the pair with
Exercise 24.3 Show that if the natural trigonometric functions have a series representation they must be
As part of the optional final project, you show these functions do satisfy the angle identities, and are periodic. This defines a natural number associated to trigonometry, much as we had the natural number associated to exponentiation.
Definition 24.5 () The natural trigonometric functions are periodic, and we define their half period as .
That is, is the smallest positive number such that and for all .
Exercise 24.4 Prove that is the smallest positive zero of .
Exercise 24.5 (Approximating with Newton’s Method) The first zero of is , so one might hope to use Newton’s method to produce an approximation for . Show the sequence starting at converges to , and use a calculator to compute the first couple terms.
This of course is not very satisfying as we had to use a calculator to find values of and ! But we know enough to approximate these values with a series expansion.
Exercise 24.6 How many terms of the series expansions of , are needed to evaluate at to within ? Use this many terms of the series expansion to approximate the terms appearing in the first two iterations of Newtons method What is your approximate value for resulting from this?
Existence of Exponentials: an Alternative Proof
Our argument above used that we had previously confirmed the existence of exponential functions, together with the Taylor Error formula to find a series representation. But as often happens, the amount of new technology we have developed along the way gives a new self-contained means of both proving the existence of exponentials, and constructing their series in one stroke! We give this alternative argument here.
The idea essentially turns some of our previous reasoning on its head: we start by looking at solutions to the equation and (1) show they satisfy the law of exponents, then (2) construct an explicit solution as a power series. First, a helpful lemma about this differential equation:
Proposition 24.7 Let be two solutions to the differential equation . Then they are constant multiples of one another.
Proof. Consider the function . Differentiating with the quotient rule,
Thus for all , which implies is a constant function, and is a constant multiple of as claimed.
Now we’re ready for the main theorem:
Theorem 24.4 Let be any differentiable function which solves and has . Then is an exponential.
Proof. Let solve and satisfy . We wish to show that for all .
So, fix an arbitrary , and consider each of these separately, defining functions and .
Differentiating,
Thus, both and satisfy the differential equation . Our previous proposition implies they are constant multiples of one another,
To find this constant we evaluate at where (using ) we have
They are equal at so the constant is :
But these two functions are precisely the left and right side of the law of exponents for . Thus their equality is equivalent to sayisfying the law of exponents for this fixed value of :
As was arbitrary, this holds for all , and is an exponential.
This proof does not establish the existence of a solution to this equation, it only says if you have a solution then its an exponential. But we may now use the theory of power series to directly construct a solution!
Proposition 24.8 The series satisfies and . Thus, it defines an exponential function.
Proof. This series converges on the entire real line via the ratio test (as checked above). Thus it defines a continuous and differentiable function on , which can be differentiated term-by-term (?thm-pseries-differentiation) to yield
Finally, plugging in zero yields , finishing the argument.