It is very common to encounter series for which it is difficult, or even virtually impossible, to determine the sum exactly. Often you try to evaluate the sum approximately by truncating it, i.e. having the index run only up to some finite \(N\text{,}\) rather than infinity. But there is no point in doing so if the series diverges 1
The authors should be a little more careful making such a blanket statement. While it is true that it is not wise to approximate a divergent series by taking \(N\) terms with \(N\) large, there are cases when one can get a very good approximation by taking \(N\) terms with \(N\) small! For example, the Taylor remainder theorem shows us that when the \(n^{\rm th}\) derivative of a function \(f(x)\) grows very quickly with \(n\text{,}\) Taylor polynomials of degree \(N\text{,}\) with \(N\) large, can give bad approximations of \(f(x)\text{,}\) while the Taylor polynomials of degree one or two can still provide very good approximations of \(f(x)\) when \(x\) is very small. As an example of this, one of the triumphs of quantum electrodynamics, namely the computation of the anomalous magnetic moment of the electron, depends on precisely this. A number of important quantities were predicted using the first few terms of divergent power series. When those quantities were measured experimentally, the predictions turned out to be incredibly accurate.
2
The field of asymptotic analysis often makes use of the first few terms of divergent series to generate approximate solutions to problems; this, along with numerical computations, is one of the most important techniques in applied mathematics. Indeed, there is a whole wonderful book (which, unfortunately, is too advanced for most Calculus 2 students) devoted to playing with divergent series called, unsurprisingly, “Divergent Series” by G.H. Hardy. This is not to be confused with the “Divergent” series by V. Roth set in a post-apocalyptic dystopian Chicago. That latter series diverges quite dramatically from mathematical topics, while the former does not have a film adaptation (yet).
. So you like to at least know if the series converges or diverges. Furthermore you would also like to know what error is introduced when you approximate \(\sum_{n=1}^\infty a_n\) by the “truncated series” \(\sum_{n=1}^Na_n\text{.}\) That’s called the truncation error. There are a number of “convergence tests” to help you with this.
Subsection3.3.1The Divergence Test
Our first test is very easy to apply, but it is also rarely useful. It just allows us to quickly reject some “trivially divergent” series. It is based on the observation that
by definition, a series \(\sum_{n=1}^\infty a_n\) converges to \(S\) when the partial sums \(S_N=\sum_{n=1}^N a_n\) converge to \(S\text{.}\)
Then, as \(N\rightarrow\infty\text{,}\) we have \(S_N\rightarrow S\) and, because \(N-1\rightarrow\infty\) too, we also have \(S_{N-1}\rightarrow S\text{.}\)
So \(a_N=S_N-S_{N-1}\rightarrow S-S=0\text{.}\)
This tells us that, if we already know that a given series \(\sum a_n\) is convergent, then the \(n^{\rm th}\) term of the series, \(a_n\text{,}\) must converge to \(0\) as \(n\) tends to infinity. In this form, the test is not so useful. However the contrapositive 3
We have discussed the contrapositive a few times in the CLP notes, but it doesn’t hurt to discuss it again here (or for the reader to quickly look up the relevant footnote in Section 1.3 of the CLP-1 text). At any rate, given a statement of the form “If A is true, then B is true” the contrapositive is “If B is not true, then A is not true”. The two statements in quotation marks are logically equivalent — if one is true, then so is the other. In the present context we have “If (\(\sum a_n\) converges) then (\(a_n\) converges to \(0\)).” The contrapositive of this statement is then “If (\(a_n\) does not converge to 0) then (\(\sum a_n\) does not converge).”
of the statement is a useful test for divergence.
Theorem3.3.1.Divergence Test.
If the sequence \(\big\{a_n\big\}_{n=1}^\infty\) fails to converge to zero as \(n\rightarrow\infty\text{,}\) then the series \(\sum_{n=1}^\infty a_n\) diverges.
So the series \(\sum_{n=1}^\infty \frac{n}{n+1}\) diverges.
Warning3.3.3.
The divergence test is a “one way test”. It tells us that if \(\lim_{n\rightarrow\infty}a_n\) is nonzero, or fails to exist, then the series \(\sum_{n=1}^\infty a_n\) diverges. But it tells us absolutely nothing when \(\lim_{n\rightarrow\infty}a_n=0\text{.}\) In particular, it is perfectly possible for a series \(\sum_{n=1}^\infty a_n\) to diverge even though \(\lim_{n\rightarrow\infty}a_n=0\text{.}\) An example is \(\sum_{n=1}^\infty \frac{1}{n}\text{.}\) We’ll show in Example 3.3.6, below, that it diverges.
Now while convergence or divergence of series like \(\sum_{n=1}^\infty \frac{1}{n}\) can be determined using some clever tricks — see the optional §3.3.9 —, it would be much better to have methods that are more systematic and rely less on being sneaky. Over the next subsections we will discuss several methods for testing series for convergence.
Note that while these tests will tell us whether or not a series converges, they do not (except in rare cases) tell us what the series adds up to. For example, the test we will see in the next subsection tells us quite immediately that the series
converges. However it does not tell us its value 4
This series converges to Apéry’s constant \(1.2020569031\dots\text{.}\) The constant is named for Roger Apéry (1916–1994) who proved that this number must be irrational. This number appears in many contexts including the following cute fact — the reciprocal of Apéry’s constant gives the probability that three positive integers, chosen at random, do not share a common prime factor.
.
Subsection3.3.2The Integral Test
In the integral test, we think of a series \(\sum_{n=1}^\infty a_n\text{,}\) that we cannot evaluate explicitly, as the area of a union of rectangles, with \(a_n\) representing the area of a rectangle of width one and height \(a_n\text{.}\) Then we compare that area with the area represented by an integral, that we can evaluate explicitly, much as we did in Theorem 1.12.17, the comparison test for improper integrals. We’ll start with a simple example, to illustrate the idea. Then we’ll move on to a formulation of the test in general.
Example3.3.4.Convergence of the harmonic series.
Visualise the terms of the harmonic series \(\sum_{n=1}^\infty\frac{1}{n}\) as a bar graph — each term is a rectangle of height \(\frac{1}{n}\) and width \(1\text{.}\) The limit of the series is then the limiting area of this union of rectangles. Consider the sketch on the left below.
It shows that the area of the shaded columns, \(\sum_{n=1}^4\frac{1}{n}\text{,}\) is bigger than the area under the curve \(y=\frac{1}{x}\) with \(1\le x\le 5\text{.}\) That is
That is the area under the curve diverges to \(+\infty\) and so the area represented by the columns must also diverge to \(+\infty\text{.}\)
It should be clear that the above argument can be quite easily generalised. For example the same argument holds mutatis mutandis 5
Latin for “Once the necessary changes are made”. This phrase still gets used a little, but these days mathematicians tend to write something equivalent in English. Indeed, English is pretty much the lingua franca for mathematical publishing. Quidquid erit.
Let \(I\) be any fixed integer with \(I \gt N_0\text{.}\) Then
\(\sum_{n=1}^\infty a_n\) converges if and only if \(\sum_{n=I}^\infty a_n\) converges — removing a fixed finite number of terms from a series cannot impact whether or not it converges.
Since \(a_n\ge 0\) for all \(n\ge I \gt N_0\text{,}\) the sequence of partial sums \(s_\ell=\sum_{n=I}^\ell a_n\) obeys \(s_{\ell+1} = s_\ell+a_{n+1} \ge s_\ell\text{.}\) That is, \(s_\ell\) increases as \(\ell\) increases.
So \(\big\{s_\ell\big\}\) must either converge to some finite number or increase to infinity. That is, either \(\sum_{n=I}^\infty a_n\) converges to a finite number or it is \(+\infty\text{.}\)
Look at the figure above. The shaded area in the figure is \(\sum_{n=I}^\infty a_n\) because
the first shaded rectangle has height \(a_I\) and width \(1\text{,}\) and hence area \(a_I\) and
the second shaded rectangle has height \(a_{I+1}\) and width \(1\text{,}\) and hence area \(a_{I+1}\text{,}\) and so on
This shaded area is smaller than the area under the curve \(y=f(x)\) for \(I-1\le x \lt \infty\text{.}\) So
and, if the integral is finite, the sum \(\sum_{n=I}^\infty a_n\) is finite too. Furthermore, the desired bound on the truncation error is just the special case of this inequality with \(I=N+1\text{:}\)
and, if the integral is infinite, the sum \(\sum_{n=I}^\infty a_n\) is infinite too.
Now that we have the integral test, it is straightforward to determine for which values of \(p\) the series 6
This series, viewed as a function of \(p\text{,}\) is called the Riemann zeta function, \(\zeta(p)\text{,}\) or the Euler-Riemann zeta function. It is extremely important because of its connections to prime numbers (among many other things). Indeed Euler proved that \(\zeta(p) = \sum_{n=1}^\infty \frac{1}{n^p}
= \prod_{\text{P prime}} \left(1 - {\rm P}^{-p} \right)^{-1}
\text{.}\) Riemann showed the connections between the zeros of this function (over complex numbers \(p\)) and the distribution of prime numbers. Arguably the most famous unsolved problem in mathematics, the Riemann hypothesis, concerns the locations of zeros of this function.
Let \(p \gt 0\text{.}\) We’ll now use the integral test to determine whether or not the series \(\sum_{n=1}^\infty\frac{1}{n^p}\) (which is sometimes called the \(p\)-series) converges.
To do so, we need a function \(f(x)\) that obeys \(f(n)=a_n=\frac{1}{n^p}\) for all \(n\) bigger than some \(N_0\text{.}\) Certainly \(f(x)=\frac{1}{x^p}\) obeys \(f(n)=\frac{1}{n^p}\) for all \(n\ge 1\text{.}\) So let’s pick this \(f\) and try \(N_0=1\text{.}\) (We can always increase \(N_0\) later if we need to.)
This function also obeys the other two conditions of Theorem 3.3.5:
\(f(x) \gt 0\) for all \(x\ge N_0=1\) and
\(f(x)\) decreases as \(x\) increases because \(f'(x)=-p\frac{1}{x^{p+1}} \lt 0\) for all \(x\ge N_0=1\text{.}\)
So the integral test tells us that the series \(\sum_{n=1}^\infty\frac{1}{n^p}\) converges if and only if the integral \(\int_1^\infty\frac{\dee{x}}{x^p}\) converges.
We have already seen, in Example 1.12.8, that the integral \(\int_1^\infty\frac{\dee{x}}{x^p}\) converges if and only if \(p \gt 1\text{.}\)
So we conclude that \(\sum_{n=1}^\infty\frac{1}{n^p}\) converges if and only if \(p \gt 1\text{.}\) This is sometimes called the \(p\)-test.
In particular, the series \(\sum_{n=1}^\infty\frac{1}{n}\text{,}\) which is called the harmonic series, has \(p=1\) and so diverges. As we add more and more terms of this series together, the terms we add, namely \(\frac{1}{n}\text{,}\) get smaller and smaller and tend to zero, but they tend to zero so slowly that the full sum is still infinite.
On the other hand, the series \(\sum_{n=1}^\infty\frac{1}{n^{1.000001}}\) has \(p = 1.000001 \gt 1\) and so converges. This time as we add more and more terms of this series together, the terms we add, namely \(\frac{1}{n^{1.000001}}\text{,}\) tend to zero (just) fast enough that the full sum is finite. Mind you, for this example, the convergence takes place very slowly — you have to take a huge number of terms to get a decent approximation to the full sum. If we approximate \(\sum_{n=1}^\infty\frac{1}{n^{1.000001}}\) by the truncated series \(\sum_{n=1}^N\frac{1}{n^{1.000001}}\text{,}\) we make an error of at most
This does tend to zero as \(N\rightarrow\infty\text{,}\) but really slowly.
We now know that the dividing line between convergence and divergence of \(\sum_{n=1}^\infty\frac{1}{n^p}\) occurs at \(p=1\text{.}\) We can dig a little deeper and ask ourselves how much more quickly than \(\frac{1}{n}\) the \(n^{\rm th}\) term needs to shrink in order for the series to converge. We know that for large \(x\text{,}\) the function \(\log x\) is smaller than \(x^a\) for any positive \(a\) — you can convince yourself of this with a quick application of L’Hôpital’s rule. So it is not unreasonable to ask whether the series
Let \(p \gt 0\text{.}\) We’ll now use the integral test to determine whether or not the series \(\sum\limits_{n=2}^\infty\frac{1}{n(\log n)^p}\) converges.
As in the last example, we start by choosing a function that obeys \(f(n)=a_n=\frac{1}{n(\log n)^p}\) for all \(n\) bigger than some \(N_0\text{.}\) Certainly \(f(x)=\frac{1}{x(\log x)^p}\) obeys \(f(n)=\frac{1}{n(\log n)^p}\) for all \(n\ge 2\text{.}\) So let’s use that \(f\) and try \(N_0=2\text{.}\)
Now let’s check the other two conditions of Theorem 3.3.5:
Both \(x\) and \(\log x\) are positive for all \(x \gt 1\text{,}\) so \(f(x) \gt 0\) for all \(x\ge N_0=2\text{.}\)
As \(x\) increases both \(x\) and \(\log x\) increase and so \(x(\log x)^p\) increases and \(f(x)\) decreases.
So the integral test tells us that the series \(\sum\limits_{n=2}^\infty\frac{1}{n(\log n)^p}\) converges if and only if the integral \(\int_2^\infty\frac{\dee{x}}{x (\log x)^p}\) converges.
To test the convergence of the integral, we make the substitution \(u=\log x\text{,}\)\(\dee{u}=\frac{\dee{x}}{x}\text{.}\)
We already know that the integral \(\int_1^\infty\frac{\dee{u}}{u^p}\text{,}\) and hence the integral \(\int_2^R \frac{\dee{x}}{x (\log x)^p}\text{,}\) converges if and only if \(p \gt 1\text{.}\)
So we conclude that \(\sum\limits_{n=2}^\infty\frac{1}{n(\log n)^p}\) converges if and only if \(p \gt 1\text{.}\)
Subsection3.3.3The Comparison Test
Our next convergence test is the comparison test. It is much like the comparison test for improper integrals (see Theorem 1.12.17) and is true for much the same reasons. The rough idea is quite simple. A sum of larger terms must be bigger than a sum of smaller terms. So if we know the big sum converges, then the small sum must converge too. On the other hand, if we know the small sum diverges, then the big sum must also diverge. Formalising this idea gives the following theorem.
Theorem3.3.8.The Comparison Test.
Let \(N_0\) be a natural number and let \(K \gt 0\text{.}\)
If \(|a_n|\le K c_n\) for all \(n\ge N_0\) and \(\sum\limits_{n=0}^\infty c_n\) converges, then \(\sum\limits_{n=0}^\infty a_n\) converges.
If \(a_n\ge K d_n\ge0 \) for all \(n\ge N_0\) and \(\sum\limits_{n=0}^\infty d_n\) diverges, then \(\sum\limits_{n=0}^\infty a_n\) diverges.
“Proof”.
We will not prove this theorem here. We’ll just observe that it is very reasonable. That’s why there are quotation marks around “Proof”. For an actual proof see the optional section 3.3.10.
If \(\sum\limits_{n=0}^\infty c_n\) converges to a finite number and if the terms in \(\sum\limits_{n=0}^\infty a_n\) are smaller than the terms in \(\sum\limits_{n=0}^\infty c_n\text{,}\) then it is no surprise that \(\sum\limits_{n=0}^\infty a_n\) converges too.
If \(\sum\limits_{n=0}^\infty d_n\) diverges (i.e. adds up to \(\infty\)) and if the terms in \(\sum\limits_{n=0}^\infty a_n\) are larger than the terms in \(\sum\limits_{n=0}^\infty d_n\text{,}\) then of course \(\sum\limits_{n=0}^\infty a_n\) adds up to \(\infty\text{,}\) and so diverges, too.
The comparison test for series is also used in much the same way as is the comparison test for improper integrals. Of course, one needs a good series to compare against, and often the series \(\sum n^{-p}\) (from Example 3.3.6), for some \(p \gt 0\text{,}\) turns out to be just what is needed.
We could determine whether or not the series \(\sum_{n=1}^\infty\frac{1}{n^2+2n+3}\) converges by applying the integral test. But it is not worth the effort 8
Go back and quickly scan Theorem 3.3.5; to apply it we need to show that \(\frac{1} {n^2+2n+3}\) is positive and decreasing (it is), and then we need to integrate \(\int \frac{1}{x^2+2x+3}\dee{x}\text{.}\) To do that we reread the notes on partial fractions, then rewrite \(x^2+2x+3
= (x+1)^2+2\) and so \(\int_1^\infty \frac{1}{x^2+2x+3}\dee{x}
= \int_1^\infty \frac{1}{(x+1)^2+2}\dee{x} \cdots\) and then arctangent appears, etc etc. Urgh. Okay — let’s go back to the text now and see how to avoid this.
. Whether or not any series converges is determined by the behaviour of the summand 9
To understand this consider any series \(\sum_{n=1}^\infty a_n\text{.}\) We can always cut such a series into two parts — pick some huge number like \(10^6\text{.}\) Then \(\sum_{n=1}^\infty a_n = \sum_{n=1}^{10^6} a_n + \sum_{n=10^6+1}^\infty a_n
\text{.}\) The first sum, though it could be humongous, is finite. So the left hand side, \(\sum_{n=1}^\infty a_n\text{,}\) is a well-defined finite number if and only if \(\sum_{n=10^6+1}^\infty a_n\text{,}\) is a well-defined finite number. The convergence or divergence of the series is determined by the second sum, which only contains \(a_n\) for “large” \(n\text{.}\)
for very large \(n\text{.}\) So the first step in tackling such a problem is to develop some intuition about the behaviour of \(a_n\) when \(n\) is very large.
Step 1: Develop intuition. In this case, when \(n\) is very large 10
The symbol “\(\gg\)” means “much larger than”. Similarly, the symbol “\(\ll\)” means “much less than”. Good shorthand symbols can be quite expressive.
\(n^2\gg 2n \gg 3\) so that \(\frac{1}{n^2+2n+3}\approx\frac{1}{n^2}\text{.}\) We already know, from Example 3.3.6, that \(\sum_{n=1}^\infty\frac{1}{n^p}\) converges if and only if \(p \gt 1\text{.}\) So \(\sum_{n=1}^\infty\frac{1}{n^2}\text{,}\) which has \(p=2\text{,}\) converges, and we would expect that \(\sum_{n=1}^\infty\frac{1}{n^2+2n+3}\) converges too.
Step 2: Verify intuition. We can use the comparison test to confirm that this is indeed the case. For any \(n\ge 1\text{,}\)\(n^2+2n+3 \gt n^2\text{,}\) so that \(\frac{1}{n^2+2n+3}\le\frac{1}{n^2}\text{.}\) So the comparison test, Theorem 3.3.8, with \(a_n=\frac{1}{n^2+2n+3}\) and \(c_n=\frac{1}{n^2}\text{,}\) tells us that \(\sum_{n=1}^\infty\frac{1}{n^2+2n+3}\) converges.
Of course the previous example was “rigged” to give an easy application of the comparison test. It is often relatively easy, using arguments like those in Example 3.3.9, to find a “simple” series \(\sum_{n=1}^\infty b_n\) with \(b_n\) almost the same as \(a_n\) when \(n\) is large. However it is pretty rare that \(a_n\le b_n\) for all \(n\text{.}\) It is much more common that \(a_n\le K b_n\) for some constant \(K\text{.}\) This is enough to allow application of the comparison test. Here is an example.
We already know from Example 3.3.6, with \(p=2\text{,}\) that \(\sum_{n=1}^\infty\frac{1}{n^2}\) converges, so we would expect that \(\sum_{n=1}^\infty\frac{n+\cos n}{n^3-\frac{1}{3}}\) converges too.
Step 2: Verify intuition. We can use the comparison test to confirm that this is indeed the case. To do so we need to find a constant \(K\) such that \(|a_n|= \frac{|n+\cos n|}{n^3-1/3}=\frac{n+\cos n}{n^3-1/3}\) is smaller than \(\frac{K}{n^2}\) for all \(n\text{.}\) A good way 11
This is very similar to how we computed limits at infinity way way back near the beginning of CLP-1.
to do that is to factor the dominant term (in this case \(n\)) out of the numerator and also factor the dominant term (in this case \(n^3\)) out of the denominator.
So now we need to find a constant \(K\) such that \(\frac{1+\frac{(\cos n)}{n}}{1-\frac{1}{3n^3}}\) is smaller than \(K\) for all \(n\ge 1\text{.}\)
First consider the numerator \(1+(\cos n)\frac{1}{n}\text{.}\) For all \(n\ge 1\)
\(\frac{1}{n}\le 1\) and
\(\displaystyle |\cos n|\le 1\)
So the numerator \(1+(\cos n)\frac{1}{n}\) is always smaller than \(1+(1)\frac{1}{1}=2\text{.}\)
Next consider the denominator \(1-\frac{1}{3n^3}\text{.}\)
When \(n\ge 1\text{,}\)\(\frac{1}{3n^3}\) lies between \(\frac{1}{3}\) and \(0\) so that
\(1-\frac{1}{3n^3}\) is between \(\frac{2}{3}\) and \(1\) and consequently
\(\frac{1}{1-\frac{1}{3n^3}}\) is between \(\frac{3}{2}\) and \(1\text{.}\)
As the numerator \(1+(\cos n)\frac{1}{n}\) is always smaller than \(2\) and \(\frac{1}{1-\frac{1}{3n^3}}\) is always smaller than \(\frac{3}{2}\text{,}\) the fraction
and, since we know \(\sum_{n=1}^\infty n^{-2}\) converges, the comparison test tells us that \(\sum_{n=1}^\infty\frac{n+\cos n}{n^3-1/3}\) converges.
The last example was actually a relatively simple application of the comparison theorem — finding a suitable constant \(K\) can be really tedious 12
Really, really tedious. And you thought some of those partial fractions computations were bad …
. Fortunately, there is a variant of the comparison test that completely eliminates the need to explicitly find \(K\text{.}\)
The idea behind this isn’t too complicated. We have already seen that the convergence or divergence of a series depends not on its first few terms, but just on what happens when \(n\) is really large. Consequently, if we can work out how the series terms behave for really big \(n\) then we can work out if the series converges. So instead of comparing the terms of our series for all \(n\text{,}\) just compare them when \(n\) is big.
Theorem3.3.11.Limit Comparison Theorem.
Let \(\sum_{n=1}^\infty a_n\) and \(\sum_{n=1}^\infty b_n\) be two series with \(b_n \gt 0\) for all \(n\text{.}\) Assume that
If \(\sum_{n=1}^\infty b_n\) converges, then \(\sum_{n=1}^\infty a_n\) converges too.
If \(L\ne 0\) and \(\sum_{n=1}^\infty b_n\) diverges, then \(\sum_{n=1}^\infty a_n\) diverges too.
In particular, if \(L\ne 0\text{,}\) then \(\sum_{n=1}^\infty a_n\) converges if and only if \(\sum_{n=1}^\infty b_n\) converges.
Proof.
(a) Because we are told that \(\lim_{n\rightarrow\infty}\frac{a_n}{b_n}=L \text{,}\) we know that,
when \(n\) is large, \(\frac{a_n}{b_n}\) is very close to \(L\text{,}\) so that \(\Big|\frac{a_n}{b_n}\Big|\) is very close to \(|L|\text{.}\)
In particular, there is some natural number \(N_0\) so that \(\Big|\frac{a_n}{b_n}\Big|\le |L|+1\text{,}\) for all \(n\ge N_0\text{,}\) and hence
\(|a_n|\le Kb_n\) with \(K=|L|+1\text{,}\) for all \(n\ge N_0\text{.}\)
The comparison Theorem 3.3.8 now implies that \(\sum_{n=1}^\infty a_n\) converges.
(b) Let’s suppose that \(L \gt 0\text{.}\) (If \(L \lt 0\text{,}\) just replace \(a_n\) with \(-a_n\text{.}\)) Because we are told that \(\lim_{n\rightarrow\infty}\frac{a_n}{b_n}=L \text{,}\) we know that,
when \(n\) is large, \(\frac{a_n}{b_n}\) is very close to \(L\text{.}\)
In particular, there is some natural number \(N\) so that \(\frac{a_n}{b_n}\ge \frac{L}{2}\text{,}\) and hence
\(a_n\ge Kb_n\) with \(K=\frac{L}{2} \gt 0\text{,}\) for all \(n\ge N\text{.}\)
The comparison Theorem 3.3.8 now implies that \(\sum_{n=1}^\infty a_n\) diverges.
The next two examples illustrate how much of an improvement the above theorem is over the straight comparison test (though of course, we needed the comparison test to develop the limit comparison test).
Set \(a_n= \frac{\sqrt{n+1}}{n^2-2n+3}\text{.}\) We first try to develop some intuition about the behaviour of \(a_n\) for large \(n\) and then we confirm that our intuition was correct.
Step 1: Develop intuition. When \(n\gg 1\text{,}\) the numerator \(\sqrt{n+1}\approx \sqrt{n}\text{,}\) and the denominator \(n^2-2n+3\approx n^2\) so that \(a_n\approx \frac{\sqrt{n}}{n^2}=\frac{1}{n^{3/2}}\) and it looks like our series should converge by Example 3.3.6 with \(p=\frac{3}{2}\text{.}\)
Step 2: Verify intuition. To confirm our intuition we set \(b_n=\frac{1}{n^{3/2}}\) and compute the limit
We already know that the series \(\sum_{n=1}^\infty b_n =\sum_{n=1}^\infty\frac{1}{n^{3/2}}\) converges by Example 3.3.6 with \(p=\frac{3}{2}\text{.}\) So our series converges by the limit comparison test, Theorem 3.3.11.
The denominator is quite a bit more tricky, since we need a lower bound, rather than an upper bound, and we cannot just write \(|n^2-2n+3| \ge n^2\text{,}\) which is false. Instead we have to make a more careful argument. In particular, we’d like to find \(N_0\) and \(K'\) so that \(n^2-2n+3\ge K'n^2\text{,}\) i.e. \(\frac{1}{n^2-2n+3}\le\frac{1}{K'n^2}\) for all \(n \geq N_0\text{.}\) For \(n\ge 4\text{,}\) we have \(2n = \frac{1}{2} 4n\le \frac{1}{2}n\cdot n=\frac{1}{2}n^2\text{.}\) So for \(n\ge 4\text{,}\)
and the comparison test then tells us that our series converges. It is pretty clear that the approach of Example 3.3.12 was much more straightforward.
Subsection3.3.4The Alternating Series Test
When the signs of successive terms in a series alternate between \(+\) and \(-\text{,}\) like for example in \(\ 1-\frac{1}{2} +\frac{1}{3}-\frac{1}{4}+ \cdots\ \text{,}\) the series is called an alternating series. More generally, the series
is alternating if every \(A_n\ge 0\text{.}\) Often (but not always) the terms in alternating series get successively smaller. That is, then \(A_1\ge A_2 \ge A_3 \ge \cdots\text{.}\) In this case:
The first partial sum is \(S_1=A_1\text{.}\)
The second partial sum, \(S_2=A_1-A_2\text{,}\) is smaller than \(S_1\) by \(A_2\text{.}\)
The third partial sum, \(S_3=S_2+A_3\text{,}\) is bigger than \(S_2\) by \(A_3\text{,}\) but because \(A_3\le A_2\text{,}\)\(S_3\) remains smaller than \(S_1\text{.}\) See the figure below.
The fourth partial sum, \(S_4=S_3-A_4\text{,}\) is smaller than \(S_3\) by \(A_4\text{,}\) but because \(A_4\le A_3\text{,}\)\(S_4\) remains bigger than \(S_2\text{.}\) Again, see the figure below.
And so on.
So the successive partial sums oscillate, but with ever decreasing amplitude. If, in addition, \(A_n\) tends to \(0\) as \(n\) tends to \(\infty\text{,}\) the amplitude of oscillation tends to zero and the sequence \(S_1\text{,}\)\(S_2\text{,}\)\(S_3\text{,}\)\(\cdots\) converges to some limit \(S\text{.}\)
This is illustrated in the figure
Here is a convergence test for alternating series that exploits this structure, and that is really easy to apply.
Theorem3.3.14.Alternating Series Test.
Let \(\big\{A_n\big\}_{n=1}^\infty\) be a sequence of real numbers that obeys
\(A_n\ge 0\) for all \(n\ge 1\) and
\(A_{n+1}\le A_n\) for all \(n\ge 1\) (i.e. the sequence is monotone decreasing) and
converges and, for each natural number \(N\text{,}\)\(S-S_N\) is between \(0\) and (the first dropped term) \((-1)^N A_{N+1}\text{.}\) Here \(S_N\) is, as previously, the \(N^{\rm th}\) partial sum \(\sum\limits_{n=1}^N (-1)^{n-1} A_n\text{.}\)
“Proof”.
We shall only give part of the proof here. For the rest of the proof see the optional section 3.3.10. We shall fix any natural number \(N\) and concentrate on the last statement, which gives a bound on the truncation error (which is the error introduced when you approximate the full series by the partial sum \(S_N\))
This tells us that \((-1)^N S_{N,\ell}\le A_{N+1}\) for all for all \(\ell \gt N+1\text{,}\) both even and odd.
So we now know that \(S_{N,\ell}\) lies between its first term, \((-1)^NA_{N+1}\text{,}\) and \(0\) for all \(\ell \gt N+1\text{.}\) While we are not going to prove it here (see the optional section 3.3.10), this implies that, since \(A_{N+1}\rightarrow 0\) as \(N\rightarrow\infty\text{,}\) the series converges and that
Example3.3.15.Convergence of the alternating harmonic series.
We have already seen, in Example 3.3.6, that the harmonic series \(\sum_{n=1}^\infty\frac{1}{n}\) diverges. On the other hand, the series \(\sum_{n=1}^\infty(-1)^{n-1}\frac{1}{n}\) converges by the alternating series test with \(A_n=\frac{1}{n}\text{.}\) Note that
\(A_n=\frac{1}{n}\ge 0\) for all \(n\ge 1\text{,}\) so that \(\sum_{n=1}^\infty(-1)^{n-1}\frac{1}{n}\) really is an alternating series, and
\(A_n=\frac{1}{n}\) decreases as \(n\) increases, and
The alternating series test tells us that, for any natural number \(N\text{,}\) the error that we make when we approximate \(\frac{1}{e}\) by the partial sum \(S_N= \sum_{n=0}^N\frac{(-1)^n}{n!}\) has magnitude no larger than \(\frac{1}{(N+1)!}\text{.}\) This tends to zero spectacularly quickly as \(N\) increases, simply because \((N+1)!\) increases spectacularly quickly as \(N\) increases 13
The interested reader may wish to check out “Stirling’s approximation”, which says that \(n!\approx \sqrt{2\pi n}\left(\frac {n}{e}\right)^{n}\text{.}\)
. For example \(20!\approx 2.4\times 10^{27}\text{.}\)
Example3.3.17.Computing \(\log\frac{11}{10}\).
We will shortly see, in Example 3.5.20, that if \(-1 \lt x\le 1\text{,}\) then
Suppose that we have to compute \(\log\frac{11}{10}\) to within an accuracy of \(10^{-12}\text{.}\) Since \(\frac{11}{10}=1+\frac{1}{10}\text{,}\) we can get \(\log\frac{11}{10}\) by evaluating \(\log(1+x)\) at \(x=\frac{1}{10}\text{,}\) so that
By the alternating series test, this series converges. Also by the alternating series test, approximating \(\log\frac{11}{10}\) by throwing away all but the first \(N\) terms
The best way to do so is simply to guess — we are not going to be able to manipulate the inequality \(\frac{1}{(N+1)\times 10^{N+1}} \le \frac{1}{10^{12}}\) into the form \(N\le \cdots\text{,}\) and even if we could, it would not be worth the effort. We need to choose \(N\) so that the denominator \((N+1)\times 10^{N+1}\) is at least \(10^{12}\text{.}\) That is easy, because the denominator contains the factor \(10^{N+1}\) which is at least \(10^{12}\) whenever \(N+1\ge 12\text{,}\) i.e. whenever \(N\ge 11\text{.}\) So we will achieve an error of less than \(10^{-12}\) if we choose \(N=11\text{.}\)
This is not the smallest possible choice of \(N\text{,}\) but in practice that just doesn’t matter — your computer is not going to care whether or not you ask it to compute a few extra terms. If you really need the smallest \(N\) that obeys \(\frac{1}{(N+1)\times 10^{N+1}} \le \frac{1}{10^{12}}\text{,}\) you can next just try \(N=10\text{,}\) then \(N=9\text{,}\) and so on.
So in this problem, the smallest acceptable \(N=10\text{.}\)
Subsection3.3.5The Ratio Test
The idea behind the ratio test comes from a reexamination of the geometric series. Recall that the geometric series
\begin{gather*}
\sum_{n=0}^\infty a_n = \sum_{n=0}^\infty a r^n
\end{gather*}
converges when \(|r| \lt 1\) and diverges otherwise. So the convergence of this series is completely determined by the number \(r\text{.}\) This number is just the ratio of successive terms — that is \(r = a_{n+1}/a_n\text{.}\)
In general the ratio of successive terms of a series, \(\frac{a_{n+1}}{a_n}\text{,}\) is not constant, but depends on \(n\text{.}\) However, as we have noted above, the convergence of a series \(\sum a_n\) is determined by the behaviour of its terms when \(n\) is large. In this way, the behaviour of this ratio when \(n\) is small tells us nothing about the convergence of the series, but the limit of the ratio as \(n\to\infty\) does. This is the basis of the ratio test.
Theorem3.3.18.Ratio Test.
Let \(N\) be any positive integer and assume that \(a_n\ne 0\) for all \(n\ge N\text{.}\)
If \(\lim\limits_{n\rightarrow\infty}\Big|\frac{a_{n+1}}{a_n}\Big| = L \lt 1\text{,}\) then \(\sum\limits_{n=1}^\infty a_n\) converges.
If \(\lim\limits_{n\rightarrow\infty}\Big|\frac{a_{n+1}}{a_n}\Big| = L \gt 1\text{,}\) or \(\lim\limits_{n\rightarrow\infty}\Big|\frac{a_{n+1}}{a_n}\Big| = +\infty\text{,}\) then \(\sum\limits_{n=1}^\infty a_n\) diverges.
Warning3.3.19.
Beware that the ratio test provides absolutely no conclusion about the convergence or divergence of the series \(\sum\limits_{n=1}^\infty a_n\) if \(\lim\limits_{n\rightarrow\infty}\Big|\frac{a_{n+1}}{a_n}\Big| = 1\text{.}\) See Example 3.3.22, below.
Proof.
(a) Pick any number \(R\) obeying \(L \lt R \lt 1\text{.}\) We are assuming that \(\Big|\frac{a_{n+1}}{a_n}\Big|\) approaches \(L\) as \(n\rightarrow\infty\text{.}\) In particular there must be some natural number \(M\) so that \(\Big|\frac{a_{n+1}}{a_n}\Big|\le R\) for all \(n\ge M\text{.}\) So \(|a_{n+1}|\le R|a_n|\) for all \(n\ge M\text{.}\) In particular
for all \(\ell\ge 0\text{.}\) The series \(\sum_{\ell=0}^\infty R^\ell \,|a_M|\) is a geometric series with ratio \(R\) smaller than one in magnitude and so converges. Consequently, by the comparison test with \(a_n\) replaced by \(A_\ell = a_{n+\ell}\) and \(c_n\) replaced by \(C_\ell= R^\ell \, |a_M|\text{,}\) the series \(\sum\limits_{\ell=1}^\infty a_{M+\ell} =\sum\limits_{n=M+1}^\infty a_n\) converges. So the series \(\sum\limits_{n=1}^\infty a_n\) converges too.
(b) We are assuming that \(\Big|\frac{a_{n+1}}{a_n}\Big|\) approaches \(L \gt 1\) as \(n\rightarrow\infty\text{.}\) In particular there must be some natural number \(M \gt N\) so that \(\Big|\frac{a_{n+1}}{a_n}\Big|\ge 1\) for all \(n\ge M\text{.}\) So \(|a_{n+1}|\ge |a_n|\) for all \(n\ge M\text{.}\) That is, \(|a_n|\) increases as \(n\) increases as long as \(n\ge M\text{.}\) So \(|a_n|\ge |a_M|\) for all \(n\ge M\) and \(a_n\) cannot converge to zero as \(n\rightarrow\infty\text{.}\) So the series diverges by the divergence test.
Example3.3.20.\(\sum_{n=0}^\infty a n x^{n-1}\).
Fix any two nonzero real numbers \(a\) and \(x\text{.}\) We have already seen in Example 3.2.4 and Lemma 3.2.5 — we have just renamed \(r\) to \(x\) — that the geometric series \(\sum_{n=0}^\infty a x^n\) converges when \(|x| \lt 1\) and diverges when \(|x|\ge 1\text{.}\) We are now going to consider a new series, constructed by differentiating 14
We shall see later, in Theorem 3.5.13, that the function \(\sum_{n=0}^\infty a n x^{n-1}\) is indeed the derivative of the function \(\sum_{n=0}^\infty a x^n\text{.}\) Of course, such a statement only makes sense where these series converge — how can you differentiate a divergent series? (This is not an allusion to a popular series of dystopian novels.) Actually, there is quite a bit of interesting and useful mathematics involving divergent series, but it is well beyond the scope of this course.
each term in the geometric series \(\sum_{n=0}^\infty a x^n\text{.}\) This new series is
The ratio test now tells us that the series \(\sum_{n=0}^\infty a\, n\, x^{n-1}\) converges if \(|x| \lt 1\) and diverges if \(|x| \gt 1\text{.}\) It says nothing about the cases \(x=\pm 1\text{.}\) But in both of those cases \(a_n=a\,n\,(\pm 1)^n\) does not converge to zero as \(n\rightarrow\infty\) and the series diverges by the divergence test.
Notice that in the above example, we had to apply another convergence test in addition to the ratio test. This will be commonplace when we reach power series and Taylor series — the ratio test will tell us something like
The series converges for \(|x| \lt R\) and diverges for \(|x| \gt R\text{.}\)
Of course, we will still have to to determine what happens when \(x=+R, -R\text{.}\) To determine convergence or divergence in those cases we will need to use one of the other tests we have seen.
Once again, fix any two nonzero real numbers \(a\) and \(X\text{.}\) We again start with the geometric series \(\sum_{n=0}^\infty a x^n\) but this time we construct a new series by integrating 15
We shall also see later, in Theorem 3.5.13, that the function \(\sum_{n=0}^\infty \frac{a}{n+1} x^{n + 1}\) is indeed an antiderivative of the function \(\sum_{n=0}^\infty a x^n\text{.}\)
each term, \(a x^n\text{,}\) from \(x=0\) to \(x=X\) giving \(\frac{a}{n+1} X^{n + 1}\text{.}\) The resulting new series is
The ratio test now tells us that the series \(\sum_{n=0}^\infty \frac{a}{n+1} X^{n + 1}\) converges if \(|X| \lt 1\) and diverges if \(|X| \gt 1\text{.}\) It says nothing about the cases \(X=\pm 1\text{.}\)
which converges by the alternating series test. See Example 3.3.15.
In conclusion, the series \(\sum_{n=0}^\infty \frac{a}{n+1} X^{n + 1}\) converges if and only if \(-1\le X \lt 1\text{.}\)
The ratio test is often quite easy to apply, but one must always be careful when the limit of the ratio is \(1\text{.}\) The next example illustrates this.
Example3.3.22.\(L=1\).
In this example, we are going to see three different series that all have \(\lim_{n\rightarrow\infty}\Big|\frac{a_{n+1}}{a_n}\Big| = 1\text{.}\) One is going to diverge and the other two are going to converge.
The fact that the terms alternate here suggests that we use the alternating series test. That will show that this series converges provided \(\frac{\sqrt{n+1}}{2n+3}\) decreases as \(n\) increases. So we define the function
(which is constructed by replacing the \(n\) in \(\frac{\sqrt{n+1}}{2n+3}\) with \(t\)) and verify that \(f(t)\) is a decreasing function of \(t\text{.}\) To prove that, it suffices to show its derivative is negative when \(t\geq 1\text{:}\)
When \(t \geq 1\) this is negative and so \(f(t)\) is a decreasing function. Thus we can apply the alternating series test to show that the series converges when \(x=\frac{1}{3}\text{.}\)
Notice that when \(n\) is large, the summand is approximately \(\frac{\sqrt{n}}{2n}\) which suggests that the series will diverge by comparison with \(\sum n^{-1/2}\text{.}\) To formalise this, we can use the limit comparison theorem:
So since this ratio has a finite limit and the series \(\sum n^{-1/2}\) diverges, we know that our series also diverges.
So in summary the series converges when \(-\frac{1}{3} \lt x \leq \frac{1}{3}\) and diverges otherwise.
Subsection3.3.6Convergence Test List
We now have half a dozen convergence tests:
Divergence Test
works well when the \(n^{\mathrm{th}}\) term in the series fails to converge to zero as \(n\) tends to infinity
Alternating Series Test
works well when successive terms in the series alternate in sign
don’t forget to check that successive terms decrease in magnitude and tend to zero as \(n\) tends to infinity
Integral Test
works well when, if you substitute \(x\) for \(n\) in the \(n^{\mathrm{th}}\) term you get a function, \(f(x)\text{,}\) that you can integrate
don’t forget to check that \(f(x)\ge 0\) and that \(f(x)\) decreases as \(x\) increases
Ratio Test
works well when \(\frac{a_{n+1}}{a_n}\) simplifies enough that you can easily compute \(\lim\limits_{n\rightarrow\infty}\big|\frac{a_{n+1}}{a_n}\big|=L\)
this often happens when \(a_n\) contains powers, like \(7^n\text{,}\) or factorials, like \(n!\)
don’t forget that \(L=1\) tells you nothing about the convergence/divergence of the series
Comparison Test and Limit Comparison Test
works well when, for very large \(n\text{,}\) the \(n^{\mathrm{th}}\) term \(a_n\) is approximately the same as a simpler term \(b_n\) (see Example 3.3.10) and it is easy to determine whether or not \(\sum_{n=1}^\infty b_n\) converges
don’t forget to check that \(b_n\ge 0\)
usually the Limit Comparison Test is easier to apply than the Comparison Test
Subsection3.3.7Optional — The Leaning Tower of Books
Imagine that you are about to stack a bunch of identical books on a table. But you don’t want to just stack them exactly vertically. You want to built a “leaning tower of books” that overhangs the edge of the table as much as possible.
How big an overhang can you get? The answer to that question, which we’ll now derive, uses a series!
Let’s start by just putting book #1 on the table. It’s the red book labelled “\(B_1\)” in the figure below.
Use a horizontal \(x\)-axis with \(x=0\) corresponding to the right hand edge of the table. Imagine that we have placed book #1 so that its right hand edge overhangs the end of the table by a distance \(x_1\text{.}\)
In order for the book to not topple off of the table, we need its centre of mass to lie above the table. That is, we need the \(x\)-coordinate of the centre mass of \(B_1\text{,}\) which we shall denote \(\bar X(B_1)\text{,}\) to obey
Assuming that our books have uniform density and are of length \(L\text{,}\)\(\bar X(B_1)\) will be exactly half way between the right hand end of the book, which is at \(x=x_1\text{,}\) and the left hand end of the book, which is at \(x=x_1-L\text{.}\) So
Now let’s put books #1 and #2 on the table, with the right hand edge of book #1 at \(x=x_1\) and the right hand edge of book #2 at \(x=x_2\text{,}\) as in the figure below.
In order for book #2 to not topple off of book #1, we need the centre of mass of book #2 to lie above book #1. That is, we need the \(x\)-coordinate of the centre mass of \(B_2\text{,}\) which is \(\bar X(B_2)=x_2-\frac{L}{2}\text{,}\) to obey
Assuming that book #2 does not topple off of book #1, we still need to arrange that the pair of books does not topple off of the table. Think of the pair of books as the combined red object in the figure
In order for the combined red object to not topple off of the table, we need the centre of mass of the combined red object to lie above the table. That is, we need the \(x\)-coordinate of the centre mass of the combined red object, which we shall denote \(\bar X(B_1\cup B_2)\text{,}\) to obey
\begin{gather*}
x_2\le x_1+\frac{L}{2}\qquad\text{and}\qquad x_1+x_2\le L
\end{gather*}
In particular we may choose \(x_1\) and \(x_2\) to satisfy \(x_2 = x_1+\frac{L}{2}\) and \(x_1+x_2 = L\text{.}\) Then, substituting \(x_2 = x_1+\frac{L}{2}\) into \(x_1+x_2 = L\) gives
Before considering the general “\(n\)-book tower”, let’s now put books #1, #2 and #3 on the table, with the right hand edge of book #1 at \(x=x_1\text{,}\) the right hand edge of book #2 at \(x=x_2\text{,}\) and the right hand edge of book #3 at \(x=x_3\text{,}\) as in the figure below.
In order for book #3 to not topple off of book #2, we need the centre of mass of book #3 to lie above book #2. That is, we need the \(x\)-coordinate of the centre mass of \(B_3\text{,}\) which is \(\bar X(B_3)=x_3-\frac{L}{2}\text{,}\) to obey
Assuming that book #3 does not topple off of book #2, we still need to arrange that the pair of books, book #2 plus book #3 (the red object in the figure below), does not topple off of book #1.
In order for this combined red object to not topple off of book #1, we need the \(x\)-coordinate of its centre mass, which we denote \(\bar X(B_2\cup B_3)\text{,}\) to obey
The centre of mass of the combined red object is the weighted average of the centre of masses of \(B_2\) and \(B_3\text{.}\) As \(B_2\) and \(B_3\) have the same weight,
Assuming that book #3 does not topple off of book #2, and also that the combined book #2 plus book #3 does not topple off of book #1, we still need to arrange that the whole tower of books, book #1 plus book #2 plus book #3 (the red object in the figure below), does not topple off of the table.
In order for this combined red object to not topple off of the table, we need the \(x\)-coordinate of its centre mass, which we denote \(\bar X(B_1\cup B_2\cup B_3)\text{,}\) to obey
The centre of mass of the combined red object is the weighted average of the centre of masses of \(B_1\) and \(B_2\) and \(B_3\text{.}\) As they all have the same weight,
We are finally ready for the general “\(n\)-book tower”. Stack \(n\) books on the table, with book \(B_1\) on the bottom and book \(B_n\) at the top, and with the right hand edge of book #\(j\) at \(x=x_j\text{.}\) The same centre of mass considerations as above show that the tower survives if
Our overhang is \(x_n = \frac{L}{2}\big(1+\frac{1}{2}+\frac{1}{3}+\cdots+\frac{1}{n}\big)\text{.}\) This is \(\frac{L}{2}\) times the \(n^{\rm th}\) partial sum of the harmonic series \(\sum_{m=1}^\infty\frac{1}{m}\text{.}\) As we saw in Example 3.3.6 (the \(p\) test), the harmonic series diverges. So, as \(n\) goes to infinity \(1+\frac{1}{2}+\frac{1}{3}+\cdots+\frac{1}{n}\) also goes to infinity. We may make the overhang as large 17
At least if our table is strong enough.
as we like!
Subsection3.3.8Optional — The Root Test
There is another test that is very similar in spirit to the ratio test. It also comes from a reexamination of the geometric series
\begin{gather*}
\sum_{n=0}^\infty a_n = \sum_{n=0}^\infty a r^n
\end{gather*}
The ratio test was based on the observation that \(r\text{,}\) which largely determines whether or not the series converges, could be found by computing the ratio \(r = a_{n+1}/a_n\text{.}\) The root test is based on the observation that \(|r|\) can also be determined by looking that the \(n^{\rm th}\) root of the \(n^{\rm th}\) term with \(n\) very large:
Of course, in general, the \(n^{\rm th}\) term is not exactly \(ar^n\text{.}\) However, if for very large \(n\text{,}\) the \(n^{\rm th}\) term is approximately proportional to \(r^n\text{,}\) with \(|r|\) given by the above limit, we would expect the series to converge when \(|r| \lt 1\) and diverge when \(|r| \gt 1\text{.}\) That is indeed the case.
Theorem3.3.24.Root Test.
Assume that
\begin{equation*}
L = \lim_{n\to\infty}\root{n}\of{\big|a_n\big|}
\end{equation*}
exists or is \(+\infty\text{.}\)
If \(L \lt 1\text{,}\) then \(\sum\limits_{n=1}^\infty a_n\) converges.
If \(L \gt 1\text{,}\) or \(L=+\infty\text{,}\) then \(\sum\limits_{n=1}^\infty a_n\) diverges.
Warning3.3.25.
Beware that the root test provides absolutely no conclusion about the convergence or divergence of the series \(\sum\limits_{n=1}^\infty a_n\) if \(\lim\limits_{n\rightarrow\infty}\root{n}\of{\big|a_n\big|} = 1\text{.}\)
Proof.
(a) Pick any number \(R\) obeying \(L \lt R \lt 1\text{.}\) We are assuming that \(\root{n}\of{|a_n|}\) approaches \(L\) as \(n\rightarrow\infty\text{.}\) In particular there must be some natural number \(M\) so that \(\root{n}\of{|a_n|}\le R\) for all \(n\ge M\text{.}\) So \(|a_n|\le R^n\) for all \(n\ge M\) and the series \(\sum\limits_{n=1}^\infty a_n\) converges by comparison to the geometric series \(\sum\limits_{n=1}^\infty R^n\)
(b) We are assuming that \(\root{n}\of{|a_n|}\) approaches \(L \gt 1\) (or grows unboundedly) as \(n\rightarrow\infty\text{.}\) In particular there must be some natural number \(M\) so that \(\root{n}\of{|a_n|}\ge 1\) for all \(n\ge M\text{.}\) So \(|a_n|\ge 1\) for all \(n\ge M\) and the series diverges by the divergence test.
We have already used the ratio test, in Example 3.3.23, to show that this series converges when \(|x| \lt \frac{1}{3}\) and diverges when \(|x| \gt \frac{1}{3}\text{.}\) We’ll now use the root test to draw the same conclusions.
We’ll now show that the limit of \(\big(n+1\big)^{\frac{1}{2n}}\) as \(n\to\infty\) is exactly \(1\text{.}\) To do, so we first compute the limit of the logarithm.
and the root test also tells us that if \(3|x| \gt 1\) the series diverges, while when \(3|x| \lt 1\) the series converges.
We have done the last example once, in Example 3.3.23, using the ratio test and once, in Example 3.3.26, using the root test. It was clearly much easier to use the ratio test. Here is an example that is most easily handled by the root test.
by Example 3.7.20 in the CLP-1 text with \(a=-1\text{.}\) As the limit is strictly smaller than \(1\text{,}\) the series \(\sum_{n=1}^\infty \big(\frac{n}{n+1}\big)^{n^2}\) converges.
To draw the same conclusion using the ratio test, one would have to show that the limit of
that appeared in Warning 3.3.3, is called the Harmonic series 18
The interested reader should use their favourite search engine to read more on the link between this series and musical harmonics. You can also find interesting links between the Harmonic series and the so-called “jeep problem” and also the problem of stacking a tower of dominoes to create an overhang that does not topple over.
are called the Harmonic numbers. Though these numbers have been studied at least as far back as Pythagoras, the divergence of the series was first proved in around 1350 by Nicholas Oresme (1320-5 – 1382), though the proof was lost for many years and rediscovered by Mengoli (1626–1686) and the Bernoulli brothers (Johann 1667–1748 and Jacob 1655–1705).
Oresme’s proof is beautiful and all the more remarkable that it was produced more than 300 years before calculus was developed by Newton and Leibnitz. It starts by grouping the terms of the harmonic series carefully:
is called the Basel problem. The problem is named after the home town of Leonhard Euler, who solved it. One can use telescoping series to show that this series must converge. Notice that
Thus, as \(k\) increases, the partial sum \(S_k\) increases (the series is a sum of positive terms), but is always smaller than \(2\text{.}\) So the sequence of partial sums converges.
Mengoli posed the problem of evaluating the series exactly in 1644 and it was solved — not entirely rigorously — by Euler in 1734. A rigorous proof had to wait another 7 years. Euler used some extremely cunning observations and manipulations of the sine function to show that
Extracting the coefficient of \(x^3\) from both expansions gives the desired result. The proof of the product formula is well beyond the scope of this course. But notice that at least the values of \(x\) which make the left hand side of (✶) zero, namely \(x=n\pi\) with \(n\) integer, are exactly the same as the values of \(x\) which make the right hand side of (✶) zero 20
Knowing that the left and right hand sides of (✶) are zero for the same values of \(x\) is far from the end of the story. Two functions \(f(x)\) and \(g(x)\) having the same zeros, need not be equal. It is certainly possible that \(f(x)=g(x)*A(x)\) where \(A(x)\) is a function that is nowhere zero. The interested reader should look up the Weierstrass factorisation theorem.
.
This approach can also be used to compute \(\sum_{n=1}^\infty n^{-2p}\) for \(p=1,2,3,\cdots\) and show that they are rational multiples 21
Search-engine your way to “Riemann zeta function”.
of \(\pi^{2p}\text{.}\) The corresponding series of odd powers are significantly nastier and getting closed form expressions for them remains a famous open problem.
Subsection3.3.10Optional — Some Proofs
In this optional section we provide proofs of two convergence tests. We shall repeatedly use the fact that any sequence \(a_1\text{,}\)\(a_2\text{,}\)\(a_3\text{,}\)\(\cdots\text{,}\) of real numbers which is increasing (i.e. \(a_{n+1}\ge a_n\) for all \(n\)) and bounded (i.e. there is a constant \(M\) such that \(a_n\le M\) for all \(n\)) converges. We shall not prove this fact 22
It is one way to state a property of the real number system called “completeness”. The interested reader should use their favourite search engine to look up “completeness of the real numbers”.
.
We start with the comparison test, and then move on to the alternating series test.
Theorem3.3.28.The Comparison Test (stated again).
Let \(N_0\) be a natural number and let \(K \gt 0\text{.}\)
If \(|a_n|\le K c_n\) for all \(n\ge N_0\) and \(\sum\limits_{n=0}^\infty c_n\) converges, then \(\sum\limits_{n=0}^\infty a_n\) converges.
If \(a_n\ge K d_n\ge0 \) for all \(n\ge N_0\) and \(\sum\limits_{n=0}^\infty d_n\) diverges, then \(\sum\limits_{n=0}^\infty a_n\) diverges.
Proof.
(a) By hypothesis \(\sum_{n=0}^\infty c_n\) converges. So it suffices to prove that \(\sum_{n=0}^\infty [Kc_n-a_n]\) converges, because then, by our Arithmetic of series Theorem 3.2.9,
\begin{equation*}
\sum_{n=0}^\infty a_n = \sum_{n=0}^\infty K c_n -\sum_{n=0}^\infty [Kc_n-a_n]
\end{equation*}
will converge too. But for all \(n\ge N_0\text{,}\)\(Kc_n-a_n\ge 0\) so that, for all \(N\ge N_0\text{,}\) the partial sums
increase with \(N\text{,}\) but never gets bigger than the finite number \(\sum\limits_{n=0}^{N_0} [Kc_n-a_n] + K \sum\limits_{n=N_0+1}^\infty c_n\text{.}\) So the partial sums \(S_N\) converge as \(N\rightarrow\infty\text{.}\)
By hypothesis, \(\sum_{n=N_0+1}^N d_n\text{,}\) and hence \(S_N\text{,}\) grows without bound as \(N\rightarrow\infty\text{.}\) So \(S_N\rightarrow\infty\) as \(N\rightarrow\infty\text{.}\)
Theorem3.3.29.Alternating Series Test (stated again).
Let \(\big\{a_n\big\}_{n=1}^\infty\) be a sequence of real numbers that obeys
\(a_n\ge 0\) for all \(n\ge 1\) and
\(a_{n+1}\le a_n\) for all \(n\ge 1\) (i.e. the sequence is monotone decreasing) and
converges and, for each natural number \(N\text{,}\)\(S-S_N\) is between \(0\) and (the first dropped term) \((-1)^N a_{N+1}\text{.}\) Here \(S_N\) is, as previously, the \(N^{\rm th}\) partial sum \(\sum\limits_{n=1}^N (-1)^{n-1} a_n\text{.}\)
Proof.
Let \(2n\) be an even natural number. Then the \(2n^{\rm th}\) partial sum obeys
So the sequence \(S_2\text{,}\)\(S_4\text{,}\)\(S_6\text{,}\)\(\cdots\) of even partial sums is a bounded, increasing sequence and hence converges to some real number \(S\text{.}\) Since \(S_{2n+1} = S_{2n} +a_{2n+1}\) and \(a_{2n+1}\) converges zero as \(n\rightarrow\infty\text{,}\) the odd partial sums \(S_{2n+1}\) also converge to \(S\text{.}\) That \(S-S_N\) is between \(0\) and (the first dropped term) \((-1)^N a_{N+1}\) was already proved in §3.3.4.
Exercises3.3.11Exercises
Exercises — Stage 1
.
1.
Select the series below that diverge by the divergence test.
Suppose there is some threshold after which a person is considered old, and before which they are young.
Let Olaf be an old person, and let Yuan be a young person.
Suppose I am older than Olaf. Am I old?
Suppose I am younger than Olaf. Am I old?
Suppose I am older than Yuan. Am I young?
Suppose I am younger than Yuan. Am I young?
4.
Below are graphs of two sequences with positive terms. Assume the sequences continue as shown. Fill in the table with conclusions that can be made from the direct comparison test, if any.
if \(\sum a_n\) converges
if \(\sum a_n\) diverges
and if \(\{a_n\}\) is the red series
then \(\sum b_n\)\(\Rule{2cm}{1pt}{0pt}\)
then \(\sum b_n\)\(\Rule{2cm}{1pt}{0pt}\)
and if \(\{a_n\}\) is the blue series
then \(\sum b_n\)\(\Rule{2cm}{1pt}{0pt}\)
then \(\sum b_n\)\(\Rule{2cm}{1pt}{0pt}\)
5.
For each pair of series below, decide whether the second series is a valid comparison series to determine the convergence of the first series, using the direct comparison test and/or the limit comparison test.
\(\displaystyle\sum_{n=10}^{\infty} \frac{1}{n-1},\) compared to the divergent series \(\displaystyle\sum_{n=10}^{\infty} \frac{1}{n}.\)
\(\displaystyle\sum_{n=1}^{\infty} \frac{\sin n}{n^2+1},\) compared to the convergent series \(\displaystyle\sum_{n=1}^{\infty} \frac{1}{n^2}.\)
\(\displaystyle\sum_{n=5}^{\infty} \frac{n^3+5n+1}{n^6-2},\) compared to the convergent series \(\displaystyle\sum_{n=5}^{\infty} \frac{1}{n^3}.\)
\(\displaystyle\sum_{n=5}^{\infty} \frac{1}{\sqrt{n}},\) compared to the divergent series \(\displaystyle\sum_{n=5}^{\infty} \frac{1}{\sqrt[4]n}.\)
6.
Suppose \(a_n\) is a sequence with \(\displaystyle\lim_{n \to \infty}a_n = \frac{1}{2}\text{.}\) Does \(\displaystyle\sum_{n=7}^\infty a_n\) converge or diverge, or is it not possible to determine this from the information given? Why?
7.
What flaw renders the following reasoning invalid?
Q: Determine whether \(\displaystyle\sum_{n=1}^\infty \dfrac{\sin n}{n}\) converges or diverges.
A: First, we will evaluate \(\displaystyle\lim_{n \to \infty} \dfrac{\sin n}{n}\text{.}\)
Note also that \(\displaystyle\lim_{n \to \infty}\frac{-1}{n}=\displaystyle\lim_{n \to \infty}\frac{1}{n}=0\text{.}\)
Therefore, by the Squeeze Theorem, \(\displaystyle\lim_{n \to \infty} \dfrac{\sin n}{n}=0\) as well.
So, by the divergence test, \(\displaystyle\sum_{n=1}^\infty \dfrac{\sin n}{n}\) converges.
8.
What flaw renders the following reasoning invalid?
Q: Determine whether \(\displaystyle\sum_{n=1}^\infty \left(\sin(\pi n)+2\right)\) converges or diverges.
A: We use the integral test. Let \(f(x)=\sin(\pi x)+2\text{.}\) Note \(f(x)\) is always positive, since \(\sin(x)+2 \geq -1+2 =1\text{.}\) Also, \(f(x)\) is continuous.
By the integral test, since the integral diverges, also \(\displaystyle\sum_{n=1}^\infty\left( \sin(\pi n)+2\right)\) diverges.
9.
What flaw renders the following reasoning invalid?
Q: Determine whether the series \(\displaystyle\sum_{n=1}^\infty \dfrac{2^{n+1}n^2}{e^n+2n}\) converges or diverges.
A: We want to compare this series to the series \(\displaystyle\sum_{n=1}^\infty \dfrac{2^{n+1}}{e^n}\text{.}\) Note both this series and the series in the question have positive terms.
First, we find that \(\dfrac{2^{n+1}n^2}{e^n+2n} \gt \dfrac{2^{n+1}}{e^n}\) when \(n\) is sufficiently large. The justification for this claim is as follows:
We note that \(e^n(n^2-1) \gt n^2-1 \gt 2n\) for \(n\) sufficiently large.
Since \(e^n+2n\) and \(e^n\) are both expressions that work out to be positive for the values of \(n\) under consideration, we can divide both sides of the inequality by these terms without having to flip the inequality. So, \(\dfrac{2^{n+1}n^2}{e^n+2n} \gt \dfrac{2^{n+1}}{e^n}\text{.}\)
Now, we claim \(\displaystyle\sum_{n=1}^\infty \dfrac{2^{n+1}}{e^n}\) converges.
Note \(\displaystyle\sum_{n=1}^\infty \dfrac{2^{n+1}}{e^n}= 2\displaystyle\sum_{n=1}^\infty \dfrac{2^{n}}{e^n}= 2\displaystyle\sum_{n=1}^\infty \left(\dfrac{2}{e}\right)^n\text{.}\) This is a geometric series with \(r=\frac{2}{e}\text{.}\) Since \(2/e \lt 1\text{,}\) the series converges.
Now, by the Direct Comparison Test, we conclude that \(\displaystyle\sum_{n=1}^\infty \dfrac{2^{n+1}n^2}{e^n+2n}\) converges.
Give an example of a convergent series for which the ratio test is inconclusive.
12.
Imagine you’re taking an exam, and you momentarily forget exactly how the inequality in the ratio test works. You remember there’s a ratio, but you don’t remember which term goes on top; you remember there’s something about the limit being greater than or less than one, but you don’t remember which way implies convergence.
should mean that the sum \(\sum\limits_{n=1}^\infty a_n\)diverges (rather than converging).
13.
Give an example of a series \(\displaystyle\sum_{n=a}^\infty a_n\text{,}\) with a function \(f(x)\) such that \(f(n)=a_n\) for all whole numbers \(n\text{,}\) such that:
\(\displaystyle\int_a^\infty f(x)\,\dee{x}\) diverges, while
\(\displaystyle\sum_{n=a}^\infty a_n\) converges.
14.(✳).
Suppose that you want to use the Limit Comparison Test on the series \(\displaystyle \sum_{n=0}^{\infty} a_n\) where \(\displaystyle a_n = \frac{2^n+n}{3^n+1}\text{.}\)
Write down a sequence \(\{b_n\}\) such that \(\displaystyle \lim\limits_{n\to\infty} \frac{a_n}{b_n}\) exists and is nonzero. (You don’t have to carry out the Limit Comparison Test)
15.(✳).
Decide whether each of the following statements is true or false. If false, provide a counterexample. If true provide a brief justification.
If \(\displaystyle\lim_{n\rightarrow\infty}a_n=0\text{,}\) then \(\sum\limits_{n=1}^{\infty} a_n\) converges.
If \(\displaystyle\lim_{n\rightarrow\infty}a_n=0\text{,}\) then \(\sum\limits_{n=1}^{\infty} (-1)^{n\mathstrut} a_n\) converges.
If \(0\le a_n \le b_n\) and \(\sum\limits_{n=1}^{\infty} b_n\) diverges, then \(\sum\limits_{n=1}^{\infty} a_n\) diverges.
Exercises — Stage 2
.
16.(✳).
Does the series \(\displaystyle \sum_{n=2}^\infty \frac{n^2}{3n^2+\sqrt n}\) converge?
17.(✳).
Determine, with explanation, whether the series \(\displaystyle \sum_{n=1}^\infty \frac{5^k}{4^k+3^k}\) converges or diverges.
18.(✳).
Determine whether the series \(\displaystyle\sum_{n=0}^\infty\frac{1}{n+\frac{1}{2}}\) is convergent or divergent. If it is convergent, find its value.
19.
Does the following series converge or diverge? \(\displaystyle\sum_{k=1}^\infty\frac{1}{\sqrt{k}\sqrt{k+1}}\)
20.
Evaluate the following series, or show that it diverges: \(\displaystyle\sum_{k=30}^\infty 3(1.001)^k\text{.}\)
21.
Evaluate the following series, or show that it diverges: \(\displaystyle\sum_{n=3}^\infty \left(\frac{-1}{5}\right)^n\text{.}\)
22.
Does the following series converge or diverge? \(\displaystyle\sum_{n=7}^\infty \sin(\pi n)\)
23.
Does the following series converge or diverge? \(\displaystyle\sum_{n=7}^\infty\cos(\pi n)\)
24.
Does the following series converge or diverge? \(\displaystyle\sum_{k=1}^\infty \frac{e^k}{k!}\text{.}\)
25.
Evaluate the following series, or show that it diverges: \(\displaystyle\sum_{k=0}^\infty\frac{2^k}{3^{k+2}}\text{.}\)
26.
Does the following series converge or diverge? \(\displaystyle\sum_{n=1}^\infty\frac{n!n!}{(2n)!}\text{.}\)
27.
Does the following series converge or diverge? \(\displaystyle\sum_{n=1}^\infty\frac{n^2+1}{2n^4+n}\text{.}\)
28.(✳).
Show that the series \(\displaystyle\sum_{n=3}^\infty \frac{5}{n(\log n)^{3/2}}\) converges.
29.(✳).
Find the values of \(p\) for which the series \(\displaystyle{\sum_{n=2}^\infty \frac{1}{n(\log n)^p}}\) converges.
30.(✳).
Does \({\displaystyle\sum_{n=1}^\infty\frac{e^{-\sqrt{n}}}{\sqrt{n}}}\) converge or diverge?
31.(✳).
Use the comparison test (not the limit comparison test) to show whether the series
Determine whether the series \(\displaystyle\sum_{n=1}^\infty\frac{n^3-4}{2n^5-6n}\) is convergent or divergent.
42.(✳).
What is the smallest value of \(N\) such that the partial sum \(\displaystyle\sum_{n=1}^N\frac{(-1)^n}{n\cdot 10^n}\) approximates \(\displaystyle\sum_{n=1}^\infty\frac{(-1)^n}{n\cdot 10^n}\) within an accuracy of \(10^{-6}\text{?}\)
43.(✳).
It is known that \(\displaystyle \sum_{n=1}^\infty \frac{(-1)^{n-1}}{n^2} = \frac{\pi^2}{12}\) (you don’t have to show this). Find \(N\) so that \(S_N\text{,}\) the \(N^{\rm th}\) partial sum of the series, satisfies \(| \frac{\pi^2}{12} - S_N | \le 10^{-6}\text{.}\) Be sure to say why your method can be applied to this particular series.
44.(✳).
The series \(\displaystyle \sum_{n=1}^\infty \frac{(-1)^{n+1}}{(2n+1)^2}\) converges to some number \(S\) (you don’t have to prove this). According to the Alternating Series Estimation Theorem, what is the smallest value of \(N\) for which the \(N^{\rm th}\) partial sum of the series is at most \(\frac1{100}\) away from \(S\text{?}\) For this value of \(N\text{,}\) write out the \(N^{\rm th}\) partial sum of the series.
Exercises — Stage 3
.
A number of phenomena roughly follow a distribution called Zipf’s law. We discuss some of these in Questions 52 and 53.
45.(✳).
Determine, with explanation, whether the following series converge or diverge.
(a) Prove that \(\displaystyle \int_2^\infty\frac{x+\sin x}{1+x^2}\ \dee{x}\) diverges.
(b) Explain why you cannot conclude that \(\displaystyle\sum\limits_{n=1}^\infty \frac{n+\sin n}{1+n^2}\) diverges from part (a) and the Integral Test.
(c) Determine, with explanation, whether \(\displaystyle\sum\limits_{n=1}^\infty \frac{n+\sin n}{1+n^2}\) converges or diverges.
47.(✳).
Show that \(\displaystyle\sum\limits_{n=1}^\infty\frac{e^{-\sqrt{n}}}{\sqrt{n}}\) converges and find an interval of length \(0.05\) or less that contains its exact value.
48.(✳).
Suppose that the series \(\displaystyle\sum\limits_{n=1}^\infty a_n\) converges and that \(1 \gt a_n\ge 0\) for all \(n\text{.}\) Prove that the series \(\displaystyle\sum\limits_{n=1}^\infty \frac{a_n}{1-a_n}\) also converges.
49.(✳).
Suppose that the series \(\sum\limits_{n=0}^{\infty}(1-a_n)\) converges, where \(a_n \gt 0\) for \(n=0,1,2,3,\cdots\text{.}\) Determine whether the series \(\sum\limits_{n=0}^\infty 2^n a_n\) converges or diverges.
50.(✳).
Assume that the series \(\displaystyle\sum_{n=1}^\infty\frac{na_n-2n+1}{n+1}\) converges, where \(a_n \gt 0\) for \(n = 1, 2, \cdots\text{.}\) Is the following series
convergent? If your answer is NO, justify your answer. If your answer is YES, evaluate the sum of the series \(-\log a_1 + \sum\limits_{n=1}^\infty \log\big(\frac{a_n}{a_{n+1}}\big)\text{.}\)
51.(✳).
Prove that if \(a_n\ge 0\) for all \(n\) and if the series \(\displaystyle\sum_{n=1}^\infty a_n\) converges, then the series \(\displaystyle\sum_{n=1}^\infty a^2_n\) also converges.
52.
Suppose the frequency of word use in a language has the following pattern:
The \(n\)-th most frequently used word accounts for \(\dfrac{\alpha}{n}\) percent of the total words used.
So, in a text of 100 words, we expect the most frequently used word to appear \(\alpha\) times, while the second-most-frequently used word should appear about \(\frac{\alpha}{2}\) times, and so on.
If books written in this language use \(20,000\) distinct words, then the most commonly used word accounts for roughly what percentage of total words used?
53.
Suppose the sizes of cities in a country adhere to the following pattern: if the largest city has population \(\alpha\text{,}\) then the \(n\)-th largest city has population \(\frac{\alpha}{n}\text{.}\)
If the largest city in this country has 2 million people and the smallest city has 1 person, then the population of the entire country is \(\sum_{n=1}^{2 \times 10^6}\frac{2\times 10^6}{n}\text{.}\) (For many \(n\)’s in this sum \(\frac{2\times 10^6}{n}\) is not an integer. Ignore that.) Evaluate this sum approximately, with an error of no more than 1 million people.