Area of a circle (Part 3)

Math majors are completely comfortable with the formula A = \pi r^2 for the area of a circle. However, they often tell me that they don’t remember a proof or justification for why this formula is true. And they certainly don’t remember a justification that would be appropriate for showing geometry students.

In this series of posts, I’ll discuss several ways that the area of a circle can be found using calculus. I’ll also discuss a straightforward classroom activity by which students can discover for themselves why A = \pi r^2.green lineIf R denotes a circular region with radius a centered at the origin, then

A = \displaystyle \iint_R 1 \, dx \, dy

This double integral may be computed by converting to polar coordinates. The distance from the origin varies from r=0 to r=a, while the angle varies from \theta = 0 to \theta = 2\pi. Using the conversion dx \, dy = r \, dr \, d\theta, we see that

A = \displaystyle \int_0^{2 \pi} \int_0^a r \, dr \, d \theta

A = \displaystyle \int_0^{2\pi} \left[ \frac{r^2}{2} \right]_0^a \, d\theta

A = \displaystyle \int_0^{2\pi} \frac{a^2}{2} \, d\theta

A = \displaystyle 2 \pi \cdot \frac{a^2}{2}

A = \displaystyle \pi a^2

We note that the above proof uses the fact that calculus with trigonometric functions must be done with radians and not degrees. In other words, we had to change the range of integration to [0,2\pi] and not [0^o, 360^o].

Area of a circle (Part 2)

Math majors are completely comfortable with the formula A = \pi r^2 for the area of a circle. However, they often tell me that they don’t remember a proof or justification for why this formula is true. And they certainly don’t remember a justification that would be appropriate for showing geometry students.

In this series of posts, I’ll discuss several ways that the area of a circle can be found using calculus. I’ll also discuss a straightforward classroom activity by which students can discover for themselves why A = \pi r^2.green line

A circle centered at the origin with radius r may be viewed as the region between f(x) = -\sqrt{r^2 - x^2} and g(x) = \sqrt{r^2 - x^2}. These two functions intersect at x = r and x = -r. Therefore, the area of the circle is the integral of the difference of the two functions:

A = \displaystyle \int_{-r}^r \left[g(x) - f(x) \right] \, dx= \displaystyle \int_{-r}^r 2 \sqrt{r^2 - x^2} \, dx

This may be evaluated by using the trigonometric substitution x = r \sin \theta and changing the range of integration to \theta = -\pi/2 to \theta = \pi/2. Since dx = r \cos \theta \, d\theta, we find

A = \displaystyle \int_{-\pi/2}^{\pi/2} 2 \sqrt{r^2 - r^2 \sin^2 \theta} \, r \cos \theta d\theta

A = \displaystyle \int_{-\pi/2}^{\pi/2} 2 r^2 \cos^2 \theta d\theta

A = \displaystyle r^2 \int_{-\pi/2}^{\pi/2} (1 + \cos 2\theta) d\theta

A = \displaystyle r^2 \left[ \theta + \frac{1}{2} \sin 2\theta \right]_{-\pi/2}^{\pi/2}

A = \displaystyle r^2 \left[ \left( \displaystyle \frac{\pi}{2} + \frac{1}{2} \sin \pi \right) - \left( - \frac{\pi}{2} + \frac{1}{2} \sin (-\pi) \right) \right]

A = \pi r^2

We note that the above proof uses the fact that calculus with trigonometric functions must be done with radians and not degrees. In other words, we had to change the range of integration to [-\pi/2,\pi/2] and not [-90^o, 90^o].

Area of a circle (Part 1)

Math majors are completely comfortable with the formula A = \pi r^2 for the area of a circle. However, they often tell me that they don’t remember a proof or justification for why this formula is true. And they certainly don’t remember a justification that would be appropriate for showing geometry students.

In this series of posts, I’ll discuss several ways that the area of a circle can be found using calculus. I’ll also discuss a straightforward classroom activity by which students can discover for themselves why A = \pi r^2.green lineIn the first few weeks after a calculus class, after students are introduced to the concept of limits, the derivative is introduced for the first time… often as the slope of a tangent line to the curve. Here it is: if $y = f(x)$, then

\displaystyle \frac{dy}{dx} = y' = f'(x) = \lim_{h \to 0} \displaystyle \frac{f(x+h) - f(x)}{h}

From this definition, the first few rules of differentiation are derived in approximately the following order:

1. If f(x) = c, a constant, then \displaystyle \frac{d}{dx} (c) = 0.

2. If f(x) and g(x) are both differentiable, then (f+g)'(x) = f'(x) + g'(x).

3. If f(x) is differentiable and c is a constant, then (cf)'(x) = c f'(x).

4. If f(x) = x^n, where n is a nonnegative integer, then f'(x) = n x^{n-1}. This may be proved by at least two different techniques:

  • The binomial expansion (x+h)^n = x^n + n x^{n-1} h + \displaystyle {n \choose 2} x^{n-2} h^2 + \dots + h^n
  • The Product Rule (derived later) and mathematical induction

5. If f(x) = a_n x^n + a_{n-1} x^{n-1} + \dots + a_1 x + a_0 is a polynomial, then f'(x) = n a_n x^{n-1} + (n-1) a_{n-1} x^{n-2} + a_1. In other words, taking the derivative of a polynomial is easy.

After doing a few examples to help these concepts sink in, I’ll show the following two examples with about 3-4 minutes left in class.

Example 1. Let A(r) = \pi r^2. Notice I’ve changed the variable from x to r, but that’s OK. Does this remind you of anything? (Students answer: the area of a circle.) What’s the derivative? Remember, \pi is just a constant. So A'(r) = \pi \cdot 2r = 2\pi r. Does this remind you of anything? (Students answer: Whoa… the circumference of a circle.)

Example 2. Now let’s try V(r) = \displaystyle \frac{4}{3} \pi r^3. Does this remind you of anything? (Students answer: the volume of a sphere.) What’s the derivative? Again, \displaystyle \frac{4}{3} \pi is just a constant. So V'(r) = \displaystyle \frac{4}{3} \pi \cdot 3r^2 = 4\pi r^2. Does this remind you of anything? (Students answer: Whoa… the surface area of a sphere.)

Hmmm. That’s interesting. The derivative of the area of a circle is the circumference of the circle, and the derivative of the area of a sphere is the surface area of the sphere. I wonder why this works. Any ideas? (Students: stunned silence.)

This is what’s known on television as a cliff-hanger, and I’ll give you the answer at the start of class tomorrow. (Students groan, as they really want to know the answer immediately.)

In the spirit of a cliff-hanger, I offer the following thought bubble before presenting the answer.

green_speech_bubble

By definition, if A(r) = \pi r^2, then

A'(r) = \displaystyle \lim_{h \to 0} \frac{ A(r+h) - A(r) }{h} = 2\pi r

The numerator may be viewed as the area of the ring between concentric circles with radii r and r+h. In other words, imagine starting with a solid red disk of radius r +h and then removing a solid white disk of radius r. The picture would look something like this:

ring

Notice that the ring has a thickness of r+h -r = h. If this ring were to be “unpeeled” and flattened, it would approximately resemble a rectangle. The height of the rectangle would be h, while the length of the rectangle would be the circumference of the circle. So

A(r + h) - A(r) \approx 2 \pi r h

and we can conclude that

A'(r) = \displaystyle \lim_{h \to 0} \frac{ 2 \pi r h}{h} = 2\pi r

By the same reasoning, the derivative of the volume of a sphere ought to be the surface area of the sphere.

Pedagogically, I find that the above discussion helps reinforce the definition of a derivative at a time when students are most willing to forget about the formal definition in favor of the various rules of differentiation.

green line

In the above work, we started with the formula for the area of the circle and then confirmed that its derivative matched the expected result. However, the above logic can be used to derive the formula for the area of a circle from the formula $C(r) = 2\pi r$ for the circumference. We begin with the observation that A'(r) = C(r), as above. Therefore, by the Fundamental Theorem of Calculus,

A(r) - A(0) = \displaystyle \int_0^r C(t) \, dt

A(r) - A(0) = \displaystyle \int_0^r 2\pi t \, dt

A(r) - A(0) = \displaystyle \left[ \pi t^2 \right]_0^r

A(r) - A(0) = \pi r^2

Since the area of a circle with radius 0 is 0, we conclude that A(r) = \pi r^2.

Pedagogically, I don’t particularly recommend this approach, as I think students would find this explanation more confusing than the first approach. However, I can see that this could be useful for reinforcing the statement of the Fundamental Theorem of Calculus.

green line

By the way, the above reasoning works for a square or cube also, but with a little twist. For a square of side length s, the area is A(s) = s^2 and the perimeter is P(s) = 4s, which isn’t the derivative of A(s). The reason this didn’t work is because the side length s of a square corresponds to the diameter of a circle, not the radius of a circle.

But, if we let x denote half the side length of a square, then the above logic works out since

A(x) = s^2 = (2x)^2 = 4x^2

and

P(x) = 4s = 4(2x) = 8x

Written in terms of the half-sidelength x, we see that A'(x) = P(x).

Why do we teach students about radians?

Throughout grades K-10, students are slowly introduced to the concept of angles. They are told that there are 90 degrees in a right angle, 180 degrees in a straight angle, and a circle has 60 degrees. They are introduced to 30-60-90 and 45-45-90 right triangles. Fans of snowboarding even know the multiples of 180 degrees up to 1440 or even 1620 degrees.

Then, in Precalculus, we make students get comfortable with \pi, \displaystyle \frac{\pi}{2}, \displaystyle \frac{\pi}{3}, \displaystyle \frac{\pi}{4}, \displaystyle \frac{\pi}{6}, and multiples thereof.

We tell students that radians and degrees are just two ways of measuring angles, just like inches and centimeters are two ways of measuring the length of a line segment.

Still, students are extremely comfortable with measuring angles in degrees. They can easily visualize an angle of 75^o, but to visualize an angle of 2 radians, they inevitably need to convert to degrees first. In his book Surely You’re Joking, Mr. Feynman!, Nobel-Prize laureate Richard P. Feynman described himself as a boy:

I was never any good in sports. I was always terrified if a tennis ball would come over the fence and land near me, because I never could get it over the fence – it usually went about a radian off of where it was supposed to go.

Naturally, students wonder why we make them get comfortable with measuring angles with radians.

The short answer, appropriate for Precalculus students: Certain formulas are a little easier to write with radians as opposed to degrees, which in turn make certain formulas in calculus a lot easier.

The longer answer, which Precalculus students would not appreciate, is that radian measure is needed to make the derivatives of \sin x and \cos x look palatable.

CircularSector_1000

Source: http://mathworld.wolfram.com/CircularSector.html

1. In Precalculus, the length of a circle arc with central angle \theta in a circle with radius r is

s = r\theta

Also, the area of a circular sector with central angle \theta in a circle with radius r is

A = \displaystyle \frac{1}{2} r^2 \theta

In both of these formulas, the angle \theta must be measured in radians.

Students may complain that it’d be easy to make a formula of \theta is measured in degrees, and they’d be right:

s = \displaystyle \frac{180 r \theta}{\pi} and A = \displaystyle \frac{180}{\pi} r^2 \theta

However, getting rid of the 180/\pi makes the following computations from calculus a lot easier.

2a. Early in calculus, the limit

\displaystyle \lim_{\theta \to 0} \frac{\sin \theta}{\theta} = 1

is derived using the Sandwich Theorem (or Pinching Theorem or Squeeze Theorem). I won’t reinvent the wheel by writing out the proof, but it can be found here. The first step of the proof uses the formula for the above formula for the area of a circular sector.

2b. Using the trigonometric identity \cos 2x = 1 - 2 \sin^2 x, we replace x by \theta/2 to find

\displaystyle \lim_{\theta \to 0} \frac{1 - \cos \theta}{\theta} = \displaystyle \lim_{\theta \to 0} \frac{2\sin^2 \displaystyle \left( \frac{\theta}{2} \right)}{ \theta}

\displaystyle \lim_{\theta \to 0} \frac{1 - \cos \theta}{\theta} = \displaystyle \lim_{\theta \to 0} \sin \left( \frac{\theta}{2} \right) \cdot \frac{\sin \displaystyle \left( \frac{\theta}{2} \right)}{ \displaystyle \frac{\theta}{2}}

\displaystyle \lim_{\theta \to 0} \frac{1 - \cos \theta}{\theta} =0 \cdot 1

\displaystyle \lim_{\theta \to 0} \frac{1 - \cos \theta}{\theta} =0

3. Both of the above limits — as well as the formulas for \sin(\alpha + \beta) and \cos(\alpha + \beta) — are needed to prove that \displaystyle \frac{d}{dx} \sin x = \cos x and \displaystyle \frac{d}{dx} \cos x = -\sin x. Again, I won’t reinvent the wheel, but the proofs can be found here.

green lineSo, to make a long story short, radians are used to make the derivatives $y = \sin x$ and $y = \cos x$ easier to remember. It is logically possible to differentiate these functions using degrees instead of radians — see http://www.math.ubc.ca/~feldman/m100/sinUnits.pdf. However, possible is not the same thing as preferable, as calculus is a whole lot easier without these extra factors of \pi/180 floating around.

Why 0^0 is undefined

TI00

Here’s an explanation for why 0^0 is undefined that should be within the grasp of pre-algebra students:

Part 1.

  • What is 0^3? Of course, it’s 0.
  • What is 0^2? Again, 0.
  • What is 0^1? Again, 0.
  • What is 0^{1/2}, or \sqrt{0}? Again, 0.
  • What is 0^{1/3}, or \sqrt[3]{0}? In other words, what number, when cubed, is 0? Again, 0.
  • What is 0^{1/10}, or \sqrt[10]{0}? In other words, what number, when raised to the 10th power, is 0. Again, 0.

So as the exponent gets closer to 0, the answer remains 0. So, from this perspective, it looks like 0^0 ought to be equal to 0.

Part 2.

  • What is 3^0. Of course, it’s 1.
  • What is 2^0. Again, 1.
  • What is 1^0. Again, 1.
  • What is \left( \displaystyle \frac{1}{2} \right)^0? Again, 1
  • What is \left( \displaystyle \frac{1}{3} \right)^0. Again, 1
  • What is \left( \displaystyle \frac{1}{10} \right)^0? Again, 1

So as the base gets closer to 0, the answer remains 1. So, from this perspective, it looks like 0^0 ought to be equal to 1.

In conclusion: looking at it one way, 0^0 should be defined to be 0. From another perspective, 0^0 should be defined to be 1.

Of course, we can’t define a number to be two different things! So we’ll just say that 0^0 is undefined — just like dividing by 0 is undefined — rather than pretend that 0^0 switches between two different values.

green line

Here’s a more technical explanation about why 0^0 is an indeterminate form, using calculus.

Part 1. As before,

\displaystyle \lim_{x \to 0^+} 0^x = \lim_{x \to 0^+} 0 = 0.

The first equality is true because, inside of the limit, x is permitted to get close to 0 but cannot actually equal 0, and there’s no ambiguity about 0^x = 0 if x >0. (Naturally, 0^x is undefined if x < 0.)

The second equality is true because the limit of a constant is the constant.

Part 2. As before,

\displaystyle \lim_{x \to 0} x^0 = \lim_{x \to 0} 1 = 1.

Once again, the first equality is true because, inside of the limit, x is permitted to get close to 0 but cannot actually equal 0, and there’s no ambiguity about x^0 = 1 if x \ne 0.

As before, the answers from Parts 1 and 2 are different. But wait, there’s more…

Part 3. Here’s another way that 0^0 can be considered, just to give us a headache. Let’s evaluate

\displaystyle \lim_{x \to 0^+} x^{1/\ln x}

Clearly, the base tends to 0 as x \to 0. Also, \ln x \to \infty as x \to 0^+, so that \displaystyle \frac{1}{\ln x} \to 0 as x \to 0^+. In other words, this limit has the indeterminate form 0^0.

To evaluate this limit, let’s take a logarithm under the limit:

\displaystyle \lim_{x \to 0^+} \ln x^{1/\ln x} = \displaystyle \lim_{x \to 0^+} \frac{1}{\ln x} \cdot \ln x

\displaystyle \lim_{x \to 0^+} \ln x^{1/\ln x} = \displaystyle \lim_{x \to 0^+} 1

\displaystyle \lim_{x \to 0^+} \ln x^{1/\ln x} = 1

Therefore, without the extra logarithm,

\displaystyle \lim_{x \to 0^+} x^{1/\ln x} = e^1 = e

Part 4. It gets even better. Let k be any positive real number. By the same logic as above,

\displaystyle \lim_{x \to 0^+} x^{\ln k/\ln x} = e^{\ln k} = k

So, for any k \ge 0, we can find a function f(x) of the indeterminate form 0^0 so that \displaystyle f(x) = k.

In other words, we could justify defining 0^0 to be any nonnegative number. Clearly, it’s better instead to simply say that 0^0 is undefined.

P.S. I don’t know if it’s possible to have an indeterminate form of 0^0 where the answer is either negative or infinite. I tend to doubt it, but I’m not sure.

A surprising appearance of e

Here’s a simple probability problem that should be accessible to high school students who have learned the Multiplication Rule:

Suppose that you play the lottery every day for about 20 years. Each time you play, the chance that you win is 1 chance in 1000. What is the probability that, after playing  1000 times, you never win?

This is a straightforward application of the Multiplication Rule from probability. The chance of not winning on any one play is 0.999. Therefore, the chance of not winning 1000 consecutive times is (0.999)^{1000}, which we can approximate with a calculator.

TIlottery1

Well, that was easy enough. Now, just for the fun of it, let’s find the reciprocal of this answer.

TIlottery2

Hmmm. Two point seven one. Where have I seen that before? Hmmm… Nah, it couldn’t be that.

What if we changed the number 1000 in the above problem to 1,000,000? Then the probability would be (0.999999)^{1000000}.

TIlottery3

There’s no denying it now… it looks like the reciprocal is approximately e, so that the probability of never winning for both problems is approximately 1/e.

Why is this happening? I offer a thought bubble if you’d like to think about this before proceeding to the answer.

green_speech_bubbleThe above calculations are numerical examples that demonstrate the limit

\displaystyle \lim_{n \to \infty} \left(1 + \frac{x}{n}\right)^n = e^x

In particular, for the special case when n = -1, we find

\displaystyle \lim_{n \to \infty} \left(1 - \frac{1}{n}\right)^n = e^{-1} = \displaystyle \frac{1}{e}

The first limit can be proved using L’Hopital’s Rule. By continuity of the function f(x) = \ln x, we have

\ln \left[ \displaystyle \lim_{n \to \infty} \left(1 + \frac{x}{n}\right)^n \right] = \displaystyle \lim_{n \to \infty} \ln \left[ \left(1 + \frac{x}{n}\right)^n \right]

\ln \left[ \displaystyle \lim_{n \to \infty} \left(1 + \frac{x}{n}\right)^n \right] = \displaystyle \lim_{n \to \infty} n \ln \left(1 + \frac{x}{n}\right)

\ln \left[ \displaystyle \lim_{n \to \infty} \left(1 + \frac{x}{n}\right)^n \right] = \displaystyle \lim_{n \to \infty} \frac{ \displaystyle \ln \left(1 + \frac{x}{n}\right)}{\displaystyle \frac{1}{n}}

The right-hand side has the form \infty/\infty as n \to \infty, and so we may use L’Hopital’s rule, differentiating both the numerator and the denominator with respect to n.

\ln \left[ \displaystyle \lim_{n \to \infty} \left(1 + \frac{x}{n}\right)^n \right] = \displaystyle \lim_{n \to \infty} \frac{ \displaystyle \frac{1}{1 + \frac{x}{n}} \cdot \frac{-x}{n^2} }{\displaystyle \frac{-1}{n^2}}

\ln \left[ \displaystyle \lim_{n \to \infty} \left(1 + \frac{x}{n}\right)^n \right] = \displaystyle \lim_{n \to \infty} \displaystyle \frac{x}{1 + \frac{x}{n}}

\ln \left[ \displaystyle \lim_{n \to \infty} \left(1 + \frac{x}{n}\right)^n \right] = \displaystyle \frac{x}{1 + 0}

\ln \left[ \displaystyle \lim_{n \to \infty} \left(1 + \frac{x}{n}\right)^n \right] = x

Applying the exponential function to both sides, we conclude that

\displaystyle \lim_{n \to \infty} \left(1 + \frac{x}{n}\right)^n= e^x

green lineIn an undergraduate probability class, the problem can be viewed as a special case of a Poisson distribution approximating a binomial distribution if there’s a large number of trials and a small probability of success.

The above calculation also justifies (in Algebra II and Precalculus) how the formula for continuous compound interest A = Pe^{rt} can be derived from the formula for discrete compound interest A = P \displaystyle \left( 1 + \frac{r}{n} \right)^{nt}

All this to say, Euler knew what he was doing when he decided that e was so important that it deserved to be named.

Square roots with a calculator (Part 11)

This is the last in a series of posts about square roots and other roots, hopefully providing a deeper look at an apparently simple concept. However, in this post, we discuss how calculators are programmed to compute square roots quickly.

Today’s movie clip, therefore, is set in modern times:

So how do calculators find square roots anyway? First, we recognize that \sqrt{a} is a root of the polynomial f(x) = x^2 - a. Therefore, Newton’s method (or the Newton-Raphson method) can be used to find the root of this function. Newton’s method dictates that we begin with an initial guess x_1 and then iteratively find the next guesses using the recursively defined sequence

x_{n+1} = x_n - \displaystyle \frac{f(x_n)}{f'(x_n)}

For the case at hand, since f'(x) = 2x, we may write

x_{n+1} = x_n - \displaystyle \frac{x_n^2 -a}{2 x_n},

which reduces to

x_{n+1} = \displaystyle \frac{2x_n^2 - (x_n^2 -a)}{2 x_n} = \frac{x_n^2 + a}{2x_n} = \frac{1}{2} \left( x_n + \frac{a}{x_n} \right)

This algorithm can be programmed using C++, Python, etc.. For pedagogical purposes, however, I’ve found that a spreadsheet like Microsoft Excel is a good way to sell this to students. In the spreadsheet below, I use Excel to find \sqrt{2000}. In cell A1, I entered 1000 as a first guess for \sqrt{2000}. Notice that this is a really lousy first guess! Then, in cell A2, I typed the formula

=1/2*(A1+2000/A1)

So Excel computes

x_2 = \frac{1}{2} \left( x_1 + \displaystyle \frac{2000}{x_1} \right) = \frac{1}{2} \left( 1000 + \displaystyle \frac{2000}{1000} \right) = 501.

Then I filled down that formula into cells A3 through A16.

squareroot

Notice that this algorithm quickly converges to \sqrt{2000}, even though the initial guess was terrible. After 7 steps, the answer is only correct to 2 significant digits (45). After 8 steps, the answer is correct to 4 significant digits (44.72). On the 9th step, the answer is correct to 9 significant digits (44.7213595).

Indeed, there’s a theorem that essentially states that, when this algorithm converges, the number of correct digits basically doubles with each successive step. That’s a lot better than the methods shown at the start of this series of posts which only produced one extra digit with each step.

This algorithm works for finding kth roots as well as square roots. Since \sqrt[k]{a} is a root of f(x) = x^k - a, Newton’s method reduces to

x_{n+1} = x_n - \displaystyle \frac{x_n^k - a}{k x_n^{k-1}} = \displaystyle \frac{(k-1) x_n^k + a}{k x_n^{k-1}} = \displaystyle \frac{k-1}{k} x_k + \frac{1}{k} \cdot \frac{a}{x_n},

which reduces to the above sequence if k =2.

See also this Wikipedia page for further historical information as well as discussion about how the above recursive sequence can be obtained without calculus.

Square roots and logarithms without a calculator (Part 10)

This is the fifth in a series of posts about calculating roots without a calculator, with special consideration to how these tales can engage students more deeply with the secondary mathematics curriculum. As most students today have a hard time believing that square roots can be computed without a calculator, hopefully giving them some appreciation for their elders.

Today’s story takes us back to a time before the advent of cheap pocket calculators: 1949.

The following story comes from the chapter “Lucky Numbers” of Surely You’re Joking, Mr. Feynman!, a collection of tales by the late Nobel Prize winning physicist, Richard P. Feynman. Feynman was arguably the greatest American-born physicist — the subject of the excellent biography Genius: The Life and Science of Richard Feynman — and he had a tendency to one-up anyone who tried to one-up him. (He was also a serial philanderer, but that’s another story.) Here’s a story involving how, in the summer of 1949, he calculated \sqrt[3]{1729.03} without a calculator.

The first time I was in Brazil I was eating a noon meal at I don’t know what time — I was always in the restaurants at the wrong time — and I was the only customer in the place. I was eating rice with steak (which I loved), and there were about four waiters standing around.

A Japanese man came into the restaurant. I had seen him before, wandering around; he was trying to sell abacuses. (Note: At the time of this story, before the advent of pocket calculators, the abacus was arguably the world’s most powerful hand-held computational device.) He started to talk to the waiters, and challenged them: He said he could add numbers faster than any of them could do.

The waiters didn’t want to lose face, so they said, “Yeah, yeah. Why don’t you go over and challenge the customer over there?”

The man came over. I protested, “But I don’t speak Portuguese well!”

The waiters laughed. “The numbers are easy,” they said.

They brought me a paper and pencil.

The man asked a waiter to call out some numbers to add. He beat me hollow, because while I was writing the numbers down, he was already adding them as he went along.

I suggested that the waiter write down two identical lists of numbers and hand them to us at the same time. It didn’t make much difference. He still beat me by quite a bit.

However, the man got a little bit excited: he wanted to prove himself some more. “Multiplição!” he said.

Somebody wrote down a problem. He beat me again, but not by much, because I’m pretty good at products.

The man then made a mistake: he proposed we go on to division. What he didn’t realize was, the harder the problem, the better chance I had.

We both did a long division problem. It was a tie.

This bothered the hell out of the Japanese man, because he was apparently well trained on the abacus, and here he was almost beaten by this customer in a restaurant.

Raios cubicos!” he says with a vengeance. Cube roots! He wants to do cube roots by arithmetic. It’s hard to find a more difficult fundamental problem in arithmetic. It must have been his topnotch exercise in abacus-land.

He writes down a number on some paper— any old number— and I still remember it: 1729.03. He starts working on it, mumbling and grumbling: “Mmmmmmagmmmmbrrr”— he’s working like a demon! He’s poring away, doing this cube root.

Meanwhile I’m just sitting there.

One of the waiters says, “What are you doing?”.

I point to my head. “Thinking!” I say. I write down 12 on the paper. After a little while I’ve got 12.002.

The man with the abacus wipes the sweat off his forehead: “Twelve!” he says.

“Oh, no!” I say. “More digits! More digits!” I know that in taking a cube root by arithmetic, each new digit is even more work that the one before. It’s a hard job.

He buries himself again, grunting “Rrrrgrrrrmmmmmm …,” while I add on two more digits. He finally lifts his head to say, “12.0!”

The waiter are all excited and happy. They tell the man, “Look! He does it only by thinking, and you need an abacus! He’s got more digits!”

He was completely washed out, and left, humiliated. The waiters congratulated each other.

How did the customer beat the abacus?

The number was 1729.03. I happened to know that a cubic foot contains 1728 cubic inches, so the answer is a tiny bit more than 12. The excess, 1.03, is only one part in nearly 2000, and I had learned in calculus that for small fractions, the cube root’s excess is one-third of the number’s excess. So all I had to do is find the fraction 1/1728, and multiply by 4 (divide by 3 and multiply by 12). So I was able to pull out a whole lot of digits that way.

A few weeks later, the man came into the cocktail lounge of the hotel I was staying at. He recognized me and came over. “Tell me,” he said, “how were you able to do that cube-root problem so fast?”

I started to explain that it was an approximate method, and had to do with the percentage of error. “Suppose you had given me 28. Now the cube root of 27 is 3 …”

He picks up his abacus: zzzzzzzzzzzzzzz— “Oh yes,” he says.

I realized something: he doesn’t know numbers. With the abacus, you don’t have to memorize a lot of arithmetic combinations; all you have to do is to learn to push the little beads up and down. You don’t have to memorize 9+7=16; you just know that when you add 9, you push a ten’s bead up and pull a one’s bead down. So we’re slower at basic arithmetic, but we know numbers.

Furthermore, the whole idea of an approximate method was beyond him, even though a cubic root often cannot be computed exactly by any method. So I never could teach him how I did cube roots or explain how lucky I was that he happened to choose 1729.03.

The key part of the story, “for small fractions, the cube root’s excess is one-third of the number’s excess,” deserves some elaboration, especially since this computational trick isn’t often taught in those terms anymore. If f(x) = (1+x)^n, then f'(x) = n (1+x)^{n-1}, so that f'(0) = n. Since f(0) = 1, the equation of the tangent line to f(x) at x = 0 is

L(x) = f(0) + f'(0) \cdot (x-0) = 1 + nx.

The key observation is that, for x \approx 0, the graph of L(x) will be very close indeed to the graph of f(x). In Calculus I, this is sometimes called the linearization of f at x =a. In Calculus II, we observe that these are the first two terms in the Taylor series expansion of f about x = a.

For Feynman’s problem, n =\frac{1}{3}, so that \sqrt[3]{1+x} \approx 1 + \frac{1}{3} x if $x \approx 0$. Then $\latex \sqrt[3]{1729.03}$ can be rewritten as

\sqrt[3]{1729.03} = \sqrt[3]{1728} \sqrt[3]{ \displaystyle \frac{1729.03}{1728} }

\sqrt[3]{1729.03} = 12 \sqrt[3]{\displaystyle 1 + \frac{1.03}{1728}}

\sqrt[3]{1729.03} \approx 12 \left( 1 + \displaystyle \frac{1}{3} \times \frac{1.03}{1728} \right)

\sqrt[3]{1729.03} \approx 12 + 4 \times \displaystyle \frac{1.03}{1728}

This last equation explains the line “all I had to do is find the fraction 1/1728, and multiply by 4.” With enough patience, the first few digits of the correction can be mentally computed since

\displaystyle \frac{1.03}{500} < \frac{1.03}{432} = 4 \times \frac{1.03}{1728} < \frac{1.03}{400}

\displaystyle \frac{1.03 \times 2}{1000} < 4 \times \frac{1.03}{1728} < \frac{1.03 \times 25}{10000}

0.00206 < 4 \times \displaystyle \frac{1.03}{1728} < 0.0025075

So Feynman could determine quickly that the answer was 12.002\hbox{something}.

By the way,

\sqrt[3]{1729.03} \approx 12.00238378\dots

\hbox{Estimate} \approx 12.00238426\dots

So the linearization provides an estimate accurate to eight significant digits. Additional digits could be obtained by using the next term in the Taylor series.

green line

I have a similar story to tell. Back in 1996 or 1997, when I first moved to Texas and was making new friends, I quickly discovered that one way to get odd facial expressions out of strangers was by mentioning that I was a math professor. Occasionally, however, someone would test me to see if I really was a math professor. One guy (who is now a good friend; later, we played in the infield together on our church-league softball team) asked me to figure out \sqrt{97} without a calculator — before someone could walk to the next room and return with the calculator. After two seconds of panic, I realized that I was really lucky that he happened to pick a number close to 100. Using the same logic as above,

\sqrt{97} = \sqrt{100} \sqrt{1 - 0.03} \approx 10 \left(1 - \displaystyle \frac{0.03}{2}\right) = 9.85.

Knowing that this came from a linearization and that the tangent line to y = \sqrt{1+x} lies above the curve, I knew that this estimate was too high. But I didn’t have time to work out a correction (besides, I couldn’t remember the full Taylor series off the top of my head), so I answered/guessed 9.849, hoping that I did the arithmetic correctly. You can imagine the amazement when someone punched into the calculator to get 9.84886\dots

Taylor series without calculus

Is calculus really necessary for obtaining a Taylor series? Years ago, while perusing an old Schaum’s outline, I found a very curious formula for the area of a circular segment:

CircularSegment_1001

A = \displaystyle \frac{R^2}{2} (\theta - \sin \theta)

The thought occurred to me that \theta was the first term in the Taylor series expansion of \sin \theta about \theta = 0, and perhaps there was a way to use this picture to generate the remaining terms of the Taylor series.

This insight led to a paper which was published in College Mathematics Journal: cmj38-1-058-059. To my surprise and delight, this paper was later selected for inclusion in The Calculus Collection: A Resource for AP and Beyond, which is a collection of articles from the publications of Mathematical Association of America specifically targeted toward teachers of AP Calculus.

Although not included in the article, it can be proven that this iterative method does indeed yield the successive Taylor polynomials of \sin \theta, adding one extra term with each successive step.

I carefully scaffolded these steps into a project that I twice assigned to my TAMS precalculus students. Both semesters, my students got it… and they were impressed to know the formula that their calculators use to compute \sin \theta. So I think this project is entirely within the grasp of precocious precalculus students.

green line

I personally don’t know of a straightforward way of obtaining the expansion of \cos \theta without calculus. However, once the expansion of \sin \theta is known, the expansion of \cos \theta can be surmised without calculus. To do this, we note that

\cos \theta = 1 - 2 \sin^2 \left( \displaystyle \frac{\theta}{2} \right) = 1 - 2 \left( \displaystyle \frac{\theta}{2} - \frac{(\theta/2)^3}{3!} + \frac{(\theta/2)^5}{5!} \dots \right)^2

Truncating the series after n terms and squaring — and being very careful with the necessary simplifications — yield the first n terms in the Taylor series of \cos \theta.

Reminding students about Taylor series (Part 6)

Sadly, at least at my university, Taylor series is the topic that is least retained by students years after taking Calculus II. They can remember the rules for integration and differentiation, but their command of Taylor series seems to slip through the cracks. In my opinion, the reason for this lack of retention is completely understandable from a student’s perspective: Taylor series is usually the last topic covered in a semester, and so students learn them quickly for the final and quickly forget about them as soon as the final is over.

Of course, when I need to use Taylor series in an advanced course but my students have completely forgotten this prerequisite knowledge, I have to get them up to speed as soon as possible. Here’s the sequence that I use to accomplish this task. Covering this sequence usually takes me about 30 minutes of class time.

I should emphasize that I present this sequence in an inquiry-based format: I ask leading questions of my students so that the answers of my students are driving the lecture. In other words, I don’t ask my students to simply take dictation. It’s a little hard to describe a question-and-answer format in a blog, but I’ll attempt to do this below.

In the previous posts, I described how I lead students to the definition of the Maclaurin series

f(x) = \displaystyle \sum_{k=0}^{\infty} \frac{f^{(k)}(0)}{k!} x^k,

which converges to f(x) within some radius of convergence for all functions that commonly appear in the secondary mathematics curriculum.

green line

Step 7. Let’s now turn to trigonometric functions, starting with f(x) = \sin x.

What’s f(0)? Plugging in, we find f(0) = \sin 0 = 0.

As before, we continue until we find a pattern. Next, f'(x) = \cos x, so that f'(0) = 1.

Next, f''(x) = -\sin x, so that f''(0) = 0.

Next, f'''(x) = -\cos x, so that f''(0) = -1.

No pattern yet. Let’s keep going.

Next, f^{(4)}(x) = \sin x, so that f^{(4)}(0) = 0.

Next, f^{(5)}(x) = \cos x, so that f^{(5)}(0) = 1.

Next, f^{(6)}(x) = -\sin x, so that f^{(6)}(0) = 0.

Next, f^{(7)}(x) = -\cos x, so that f^{(7)}(0) = -1.

OK, it looks like we have a pattern… albeit more awkward than the patterns for e^x and \displaystyle \frac{1}{1-x}. Plugging into the series, we find that

\displaystyle \sin x= x - \frac{x^3}{3!} + \frac{x^5}{5!} - \frac{x^7}{7!} \dots

If we stare at the pattern of terms long enough, we can write this more succinctly as

\sin x = \displaystyle \sum_{n=0}^\infty (-1)^n \frac{x^{2n+1}}{(2n+1)!}

The (-1)^n term accounts for the alternating signs (starting on positive with n=0), while the 2n+1 is needed to ensure that each exponent and factorial is odd.

Let’s see… \sin x has a Taylor expansion that only has odd exponents. In what other sense are the words “sine” and “odd” associated?

In Precalculus, a function f(x) is called odd if f(-x) = -f(x) for all numbers x. For example, f(x) = x^9 is odd since f(-x) = (-x)^9 = -x^9 since 9 is a (you guessed it) an odd number. Also, \sin(-x) = -\sin x, and so \sin x is also an odd function. So we shouldn’t be that surprised to see only odd exponents in the Taylor expansion of \sin x.

A pedagogical note: In my opinion, it’s better (for review purposes) to avoid the \displaystyle \sum notation and simply use the “dot, dot, dot” expression instead. The point of this exercise is to review a topic that’s been long forgotten so that these Taylor series can be used for other purposes. My experience is that the \displaystyle \sum adds a layer of abstraction that students don’t need to overcome for the time being.

green line

Step 8. Let’s now turn try f(x) = \cos x.

What’s f(0)? Plugging in, we find f(0) = \cos 1 = 0.

Next, f'(x) = -\sin x, so that f'(0) = 0.

Next, f''(x) = -\cos x, so that f'(0) = -1.

It looks like the same pattern of numbers as above, except shifted by one derivative. Let’s keep going.

Next, f'''(x) = \sin x, so that f'''(0) = 0.

Next, f^{(4)}(x) = \cos x, so that f^{(4)}(0) = 1.

Next, f^{(5)}(x) = -\sin x, so that f^{(5)}(0) = 0.

Next, f^{(6)}(x) = -\cos x, so that f^{(6)}(0) = -1.

OK, it looks like we have a pattern somewhat similar to that of $\sin x$, except only involving the even terms. I guess that shouldn’t be surprising since, from precalculus we know that \cos x is an even function since \cos(-x) = \cos x for all x.

Plugging into the series, we find that

\displaystyle \cos x= 1 - \frac{x^2}{2!} + \frac{x^4}{4!} - \frac{x^6}{6!} \dots

If we stare at the pattern of terms long enough, we can write this more succinctly as

\cos x = \displaystyle \sum_{n=0}^\infty (-1)^n \frac{x^{2n}}{(2n)!}

green line

As we saw with e^x, the above series converge quickest for values of x near 0. In the case of \sin x and \cos x, this may be facilitated through the use of trigonometric identities, thus accelerating convergence.

For example, the series for \cos 1000^o will converge quite slowly (after converting 1000^o into radians). However, we know that

\cos 1000^o= \cos(1000^o - 720^o) =\cos 280^o

using the periodicity of \cos x. Next, since $\latex 280^o$ is in the fourth quadrant, we can use the reference angle to find an equivalent angle in the first quadrant:

\cos 1000^o = \cos 280^o = \cos(360^o - 80^o) = \cos 80^o

Finally, using the cofunction identity \cos x = \sin(90^o - x), we find

\cos 1000^o = \cos 80^o = sin(90^o - 80^o) = \sin 10^o.

In this way, the sine or cosine of any angle can be reduced to the sine or cosine of some angle between 0^o and $45^o = \pi/4$ radians. Since \pi/4 < 1, the above power series will converge reasonably rapidly.

green line

Step 10. For the final part of this review, let’s take a second look at the Taylor series

e^x = \displaystyle 1 + x + \frac{x^2}{2} + \frac{x^3}{3} + \frac{x^4}{4} + \frac{x^5}{5} + \frac{x^6}{6} + \frac{x^7}{7} + \dots

Just to be silly — for no apparent reason whatsoever, let’s replace x by ix and see what happens:

e^{ix} = \displaystyle 1 - \frac{x^2}{2!} + \frac{x^4}{4!} - \frac{x^6}{6!} \dots + i \left[\displaystyle x - \frac{x^3}{3!} + \frac{x^5}{5!} - \frac{x^7}{7!} \dots \right]

after separating the terms that do and don’t have an i.

Hmmmm… looks familiar….

So it makes sense to define

e^{ix} = \cos x + i \sin x,

which is called Euler’s formula, thus proving an unexpected connected between e^x and the trigonometric functions.