2  Dubious Computations

We already saw, when visiting infinite processes from antiquity, that it is very easy to get confused and derive a contradiction when working with infinity. But on the other hand, infinite arguments turn out to be so useful that they are irresistible! Certain objects, like π, seem to be all but out of reach without invoking infinity somewhere, and while the lessons of the ancients implore us to be careful, more than once a good mathematician has thrown caution to the wind, in the hopes of gazing upon startling new truths.

2.1 Convergence, Concern and Contradiction

2.1.1 Madhava, Leibniz & π/4

Madhava was a Indian mathematician who discovered many infinite expressions for trigonometric functions in the 1300’s, results which today are known as Taylor Series after Brook Taylor, who worked with them in 1715. In a particularly important example, Madhava found a formula to calculate the arc length along a circle, in terms of the tangent: or phrased more geometrically, the arc of a circle contained in a triangle with base of length 1.

The first term is the product of the given sine and radius of the desired arc divided by the cosine of the arc. The succeeding terms are obtained by a process of iteration when the first term is repeatedly multiplied by the square of the sine and divided by the square of the cosine. All the terms are then divided by the odd numbers 1, 3, 5, …. The arc is obtained by adding and subtracting respectively the terms of odd rank and those of even rank.

As an equation, this gives

θ=sinθcosθ13sin2θcos2θ(sinθcosθ)+15sin2θcos2θ(sin2θcos2θsinθcosθ)+

=tanθtan3θ3+tan5θ5tan7θ7+tan9θ9

If we take the arclength π/4 (the diagonal of a square), then both the base and height of our triangle are equal to 1, and this series becomes

π4=113+1517+

This result was also derived by Leibniz (one of the founders of modern calcuous), using a method close to something you might see in Calculus II these days. It goes as follows: we know (say from the last chapter) the sum of the geometric series

n0rn=11r

Thus, substituting in r=x2 gives

n0(1)nx2n=11+x2

and the right hand side of this is the derivative of arctangent! So, anti-differentiating both sides of the equation yields

arctanx=n0(1)nx2ndx=n0(1)nx2ndx=n0(1)nx2n+12n+1

Finaly, we take this result and plug in x=1: since arctan(1)=π/4 this gives what we wanted:

π4=n0(1)n12n+1=113+1517+

This argument is completely full of steps that should make us worried:

  • Why can we substitute a variable into an infinite expression and ensure it remains valid?
  • Why is the derivative of arctan a rational function?
  • Why can we integrate an infinite expression?
  • Why can we switch the order of taking an infinte sum, and integration?
  • How do we know which values of x the resulting equation is valid for?

But beyond all of this, we should be even more worried if we try to plot the graphs of the partial sums of this supposed formula for the arctangent.

The infinite series we derived seems to match the arctangent exactly for a while, and then abruptly stop, and shoot off to infinity. Where does it stop? *Right at the point we are interested in, θ=π/4, so tan(θ)=1. So, even a study of which intervals a series converges in will not be enough here, we need a theory that is so precise, it can even tell us exactly what happens at the single point forming the boundary between order and chaos.

And perhaps, before thinking the eventual answer might simply say the series always converges at the endpoints, it turns out at the other endpoint x=1, this series itself diverges! So whatever theory we build will have to account for such messy cases.

2.1.2 Dirichlet & log2

In 1827, Dirichlet was studying the sums of infinitely many terms, thinking about the alternating harmonic series n1(1)nn+1

Like the previous example, this series naturally emerges from manipulations in calculus: beginning once more with the geometric series n0rn=11r. We substitute r=x to get a series for 1/(1+x) and then integrate term by term to produce a series for the logarithm:

log(1+x)=11+xdx=n0(1)nxn

=n0(1)nxn+1n+1=xx22+x33x44+

Finally, plugging in x=1 yields the sum of interest. It turns out not to be difficult to prove that this series does indeed approach a finite value after the addition of infinitely many terms, and a quick check adding up the first thousand terms gives an approximate value of 0.6926474305598, which is very close to log(2) as expected..

log(2)=112+1314+1516+1718+19110

What happens if we multiply both sides of this equation by 2?

2log(2)=21+2312+2513+2714+2915

We can simplify this expression a bit, by re-ordering the terms to combine similar ones:

2log(2)=(21)12+(2313)14+(2515)=112+1314+15

After simplifying, we’ve returned to exactly the same series we started with! That is, we’ve shown 2log(2)=log(2), and dividing by log(2) (which is nonzero!) we see that 2=1, a contradiction!

What does this tell us? Well, the only difference between the two equations is the order in which we add the terms. And, we get different results! This reveals perhaps the most shocking discovery of all, in our time spent doing dubious computations: infinite addition is not always commutative, even though finite addition always is.

Here’s an even more dubious-looking example where we can prove that 0=log2. First, consider the infinite sum of zeroes:

0=0+0+0+0+0+

Now, rewrite each of the zeroes as xx for some specially chosen xs:

0=(11)+(1212)+(1313)+(114)+

Now, do some re-arranging to this:

(1+121)+(13+1412)+(15+1613)+

Make sure to convince yourselves that all the same terms appear here after the rearrangement!

Simplifying this a bit shows a pattern:

(112)+(1314)+(1516)+

Which, after removing the parentheses, is the familiar series (1)nn. But this series equals log(2) (or, was it 2log2?) So, if we are to believe that arithmetic with infinite sums is valid, we reach the contradiction

0=log2

2.2 Infinite Expressions for sin(x)

The sine function (along with the other trigonometric, exponential, and logarithmic functions) differs from the common functions of early mathematics (polynomials, rational functions and roots) in that it is defined not by a formula but geometrically.

Such a definition is difficult to work with if one actually wishes to compute: for example, Archimedes after much trouble managed to calculate the exact value of sin(π/96) using a recursive doubling procedure, but he would have failed to calculate sin(π/97) - 97 is not a multiple of a power of 2, so his procedure wouldn’t apply! The search for a general formula that you could plug numbers into and compute their sine, was foundational to the arithmetization of geometry.

2.2.1 Infinite Sum of Madhava

Beyond the series for the arctangent, Madhava also found an infinite series for the sine function. The first thing that needs to be proven is that sin(x) satisfies the following integral equation: (Check this, using your calculus knowledge!) sin(θ)=θ0θ0tsin(u)dudt

This equation mentions sine on both sides, which means we can use it as a recurrence relation to find better and better approximations of the sine function.

Definition 2.1 (Integral Recurrence For sin(x).) We define a sequence of functions sn(x) recursively as follows:

sn+1(θ)=θ0θ0tsn(u)dudt

Given any starting function s0(x), applying the above produces a sequence s1(x), s2(x), s3(x), which we will use to approximate the sine function.

Example 2.1 (The Series for sin(x)) Like any recursive procedure, we need to start somewhere: so let’s begin with the simplest possible (and quite incorrect) “approximation” that s0(θ)=0. Integrating this twice still gives zero, so our first approximation is s1(θ)=θ0θ0t0dudt=θ0=θ

Now, plugging in s1=θ yields our second approximation: s2(θ)=θ0θ0tududt=θ0θu22|0tdt=θ0θt22dt=θt332|0θ=θθ33!

Repeating gives the third,

s3(θ)=θ0θ0t(uu33)dudt=θ0θ(t22t443!)dt=θ(θ332θ5543!)=θθ33!+θ55!

Carrying out this process infinitely many times yields a conjectured formula for the sine function as an infinite polynomial:

Proposition 2.1 (Madhava Infinite Sine Series) sin(θ)=θθ33!+θ55!θ77!+θ99!θ1111!+

Exercise 2.1 Find a similar recursive equation for the cosine function, and use it to derive the first four terms of its series expansion.

One big question about this procedure is why in the world should this work? We found a function that sin(x) satisfies, and then we plugged something else into that function and started iterating: what justification do we have that this should start to approach the sine? We can check after the fact that it (seems to have) worked, but this leaves us far from any understanding of what is actually going on.

2.2.2 Infinite Product of Euler

Another infinite expression for the sine function arose from thinking about the behavior of polynomials, and the relation of their formulas to their roots. As an example consider a quartic polynomial p(x) with roots at x=a,b,c,d. Then we can recover p up to a constant multiple as a product of linear factors with roots at a,b,c,d. If the yintercept is p(0)=k, we can give a fully explicit description

p(x)=k(1xa)(1xb)(1xc)(1xd)

In 17334, Euler attempted to apply this same reasoning in the infinite case to the trigonometric function sin(x). This has roots at every integer multiple of π, and so following the finite logic, should factor as a product of linear factors, one for each root. There’s a slight technical problem in directly applying the above argument, namely that sin(x) has a root at x=0, so k=0. One work-around is to consider the function sinxx. This is not actually defined at x=0, but one can prove limx0sinxx=1, and attempt to use k=1

GRAPH

Its roots agree with that of sin(x) except there is no longer one at x=0. That is, the roots are ,3π,2π,π,π,2π,3π,, and the resulting factorization is

sinxx=(1+x3π)(1+x2π)(1+xπ)(1xπ)(1x2π)(1x3π)

Euler noticed all the factors come in pairs, each of which represented a difference of squares.

(1xnπ)(1+x2nπ)=(1x2n2π2)

Not worrying about the fact that infinite multiplication may not be commutative (a worry we came to appreciate with Dirichlet, but this was after Euler’s time!), we may re-group this product pairing off terms like this, to yield

sinxx=(1x2π2)(1x222π2)(1x232π2)

Finally, we may multiply back through by x and get an infinite product expression for the sine function:

Proposition 2.2 (Euler) sinx=x(1x2π2)(1x24π2)(1x29π2)

This incredible identity is actually correct: there’s only one problem - the argument itself is wrong!

Exercise 2.2 In his argument, Euler crucially uses that if we know

  • all the zeroes of a function
  • the value of that function is 1 at x=0

then we can factor the function as an infinite polynomial in terms of its zeroes. This implies that a function is completely determined by its value at x=0 and its zeroes (because after all, once you know that information you can just write down a formula like Euler did!) This is absolutely true for all finite polynomials, but it fails spectacularly in general.

Show that this is a serious flaw in Euler’s reasoning by finding a different function that has all the same zeroes as sin(x)/x and is equal to 1 at zero (in the limit)!

Exercise 2.3 (The Wallis Product for π) In 1656 John Wallis derived a remarkably beautiful formula for π (though his argumnet was not very rigorous).

π2=2123434565678789109101112111213

Using Euler’s infinite product for sin(x) evaluated at x=π/2, give a derivation of Wallis’ formula.

2.2.3 The Basel Problem

The Italian mathematician Pietro Mengoli proposed the following problem in 1650:

Definition 2.2 (The Basel Problem) Find the exact value of the infinite sum 1+122+132+142+152+

By directly computing the first several terms of this sum one can get an estimate of the value, for instance adding up the first 1,000 terms we find 1+122+132+11,0002=1.6439345, and ading the first million terms gives

1+122+132++11,0002++11,000,0002=1.64492406

so we might feel rather confident that the final answer is somewhat close to 1.64. But the interesting math problem isn’t to approximate the answer, but rather to figure out something exact, and knowing the first few decimals here isn’t of much help.

This problem was attempted by famous mathematicians across Europe over the next 80 years, but all failed. All until a relatively unknown 28 year old Swiss mathematician named Leonhard Euler published a solution in 1734, and immediately shot to fame. (In fact, this problem is named the Basel problem after Euler’s hometown.)

Proposition 2.3 (Euler) n11n2=π26

Euler’s solution begins with two different expressions for the function sin(x)/x, which he gets from the sine’s series expansion, and his own work on the infinite product:

sinxx=1x23!+x45!x67!+x89!x1011!+=(1x2π2)(1x222π2)(1x232π2)

Because two polynomials are the same if and only if the coefficients of all their terms are equal, Euler attempts to generalize this to infinite expressions, and equate the coefficients for sin. The constant coefficient is easy - we can read it off as 1 from both the series and the product, but the quadratic term already holds a deep and surprising truth.

From the series, we can again simply read off the coefficient as 1/3!. But from the product, we need to think - after multiplying everything out, what sort of products will lead to a term with x2? Since each factor is already quadratic this is more straightforward than it sounds at first - the only way to get a quadratic term is to take one of the quadratic terms already present in a factor, and multiply it by 1 from another factor! Thus, the quadratic terms are x222π2x232π2x242π2. Setting the two coefficients equal (and dividing out the negative from each side) yields

13!=1π2+122π2+132π2+

Which quickly leads to a solution to the original problem, after multiplying by π2:

π23!=1+122+132+

Euler had done it! There are of course many dubious steps taken along the way in this argument, but calculating the numerical value,

π23!=1.64493406685

We find it to be exactly the number the series is heading towards. This gave Euler the confidence to publish, and the rest is history.

But we analysis students should be looking for potential troubles in this argument. What are some that you see?

2.2.4 Viète’s Infinite Trigonometric Identity

Viete was a French mathematician in the mid 1500s, who wrote down for the first time in Europe, an exact expression for π in 1596.

Proposition 2.4 (Viète’s formula for π) 2π=222+222+2+222+2+2+22

Viete’s original publicaiton of this formula - it predates our modern notation for square roots!

How could one derive such an incredible looking expression? One approach uses trigonometric identities…an infinite number of times! Start with the familiar function sin(x). Then we may apply the double angle identity to rewrite this as

sin(x)=2sin(x2)cos(x2)

Now we may apply the double angle identity once again to the term sin(x/2) to get

sin(x)=2sin(x2)cos(x2)=4sin(x4)cos(x4)cos(x2)

and again

sin(x)=8sin(x8)cos(x8)cos(x4)cos(x2)

and again

sin(x)=16sin(x16)cos(x16)cos(x8)cos(x4)cos(x2)

And so on….after the nth stage of this process one can re-arrange the the above into the following (completely legitimate) identity:

sinx2nsinx2n=cosx2cosx4cosx8cosx16cosx2n

Viete realized that as n gets really large, the function 2nsin(x/2n) starts to look a lot like the function x…and making this replacement in the formula as we let n go to infinity yields

Proposition 2.5 (Viète’s Trigonometric Identity) sinxx=cosx2cosx4cosx8cosx16

An incredible, infinite trigonometric identity! Of course, there’s a huge question about its derivation: are we absolutely sure we are justified in making the denominator there equal to x? But carrying on without fear, we may attempt to plug in x=π/2 to both sides, yielding

2π=cosπ4cosπ8cosπ16cosπ32

Now, we are left just to simplify the right hand side into something computable, using more trigonometric identities! We know cosπ/4 is 22, and we can evaluate the other terms iteratively using the half angle identity:

cosπ8=1+cosπ42=1+222=2+22

cosπ16=1+cosπ82=1+2+222=2+2+22

Substituting these all in gives the original product. And, while this derivation has a rather dubious step in it, the end result seems to be correct! Computing the first ten terms of this product on a computer yields 0.63662077105, wheras 2/π=0.636619772. In fact, Viete used his own formula to compute an approximation of π to nine correct decimal digits. This leaves the obvious question, Why does this argument work?

2.3 The Infinitesimal Calculus

In trying to formalize many of the above arguments, mathematicians needed to put the calculus steps on a firm footing. And this comes with a whole collection of its own issues. Arguments trying to explain in clear terms what a derivative or integral was really supposed to be often led to nonsensical steps, that cast doubt on the entire procedure. Indeed, the history of calculus is itself so full of confusion that it alone is often taken as the motivation to develop a rigorous study of analysis. Because we have already seen so many other troubles that come from the infinite, we will content ourselves with just one example here: what is a derivative?

The derivative is meant to measure the slope of the tangent line to a function. In words, this is not hard to describe. But like the sine function, this does not provide a means of computing, and we are looking for a formula. Approximate formulas are not hard to create: if f(x) is our function, and h is some small number the quantity

f(x+h)f(x)h

represents the slope of the secant line to f between x and h. For any finite size in h this is only an approximation, and so thinking of this like Archimedes did his polygons and the circle, we may decide to write down a sequence of ever better approximations:

Dn=f(x+1n)f(x)1n

and then define the derivative as the infiniteth term in this sequence. But this is just incoherent, taken at face value. If 1/n0 as n this would lead us to

f(x+0)f(x)0=00

So, something else must be going on. One way out of this would be if our sequence of approximates did not actually converge to zero - maybe there were infinitely small nonzero numbers out there waiting to be discovered. Such hypothetical numbers were called infinitesimals.

Definition 2.3 (Infinitesimal) A positive number ϵ is infinitesimal if it is smaller than 1/n for all nN.

This would resolve the problem as follows: if dx is some infinitesimal number, we could define the derivative as

D=f(x+dx)f(x)dx

But this leads to its own set of difficulties: its easy to see that if ϵ is an infinitesimal, then so is 2ϵ, or kϵ for any rational number k.

Exercise 2.4 Prove this: if ϵ is infinitesimal and kQ show kϵ is infinitesimal$.

So we can’t just say define the derivative by saying “choose some infinitesimal dx” - there are many such infinitesimals and we should be worried about which one we pick! What actually happens if we try this calculation in practice, showcases this.

Let’s attempt to differentiate x2, using some infinitesimal dx. We get

(x2)=(x+dx)2x2dx=x2+2xdx+dx2x2dx =2xdx+dx2dx=2x+dx

Here we see the derivative is not what we expected, but rather is 2x plus an infinitesimal! How do we get rid of this? One approach (used very often in the foundational works of calculus) is simply to discard any infinitesimal that remains at the end of a computation. So here, because 2x is finite in size and dx is infinitesimal, we would just discard the dx and get (x2)=2x as desired.

But this is not very sensible: when exactly are we allowed to do this? If we can discard an infinitesimal whenever its added to a finite number, shouldn’t we already have done so with the (x+dx) that showed up in the numerator? This would have led to

(x+dx)2x2dx=x2x2dx=0dx=0

So, the when we throw away the infinitesimal matters deeply to the answer we get! This does not seem right. How can we fix this? One approach that was suggested was to say that we cannot throw away infinitesimals, but that the square of an infinitesimal is so small that it is precisely zero: that way, we keep every infinitesimal but discard any higher powers. A number satisfying this property was called nilpotent as nil was another word for zero, and potency was an old term for powers (x2 would be the *second potency of x).

Definition 2.4 A number ϵ is nilpotent if ϵ0 but ϵ2=0.

If our infinitesimals were nilpotent, that would solve the problem we ran into above. Now, the calculation for the derivative of x2 would proceed as

(x+dx)2x2dx=x2+2xdx+dx2x2=2xdx+0dx=2x

But, in trying to justify just this one calculation we’ve had to invent two new types of numbers that had never occurred previously in math: we need positive numbers smaller than any rational, and we also need them (or at least some of those numbers) to square to precisely zero. Do such numbers exist?