Skip to main content

PLP An introduction to mathematical proof

Appendix B Scratchwork for Exercises

1 Sets
1.4 Exercises

1.4.1.

Scratchwork.
  1. \(A_1\) is the set of all natural numbers whose square is less than \(2\text{.}\) We see that there is only one natural number satisfying that condition, which is \(1\text{.}\)
    Thus, \(A_1=\set{1}\text{.}\)
  2. We see that this set looks very similar to \(A_1\text{.}\) But we see that it is different in that we now are looking at the set of all integers whose square is less than \(2\text{,}\) not natural numbers. In this case, we see that we have \(3\) numbers satisfying the condition, namely \(-1\text{,}\) \(0\text{,}\) and \(1\text{.}\)
    Thus, \(A_2=\set{-1,0,1}\text{.}\)
  3. \(A_3\) is the set of all natural numbers that are multiples of \(3\) and also divisors of \(216\text{.}\)
    Therefore, \(A_3=\set{ 3, 6, 9, 12, 18, 24, 27, 36, 54, 72, 108, 216}\text{.}\)
  4. \(A_4\) is the set of all integers \(x\) such that if we add \(2\) and divide the result by \(5\) we still get an integer
    So, we see that \(A_4=\set{\ldots, -12, -7, -2, 3, 8, 13, \ldots}\text{.}\)
  5. We see that \(A_5\) is the set of all elements \(a\) in \(B\) such that \(6\leq 4a+1 \lt 1\text{.}\) So, all we need to do is to check which elements of \(B\) satisfy this condition.
    Then, we see \(A_5=\set{2,3}\text{.}\)
  6. We see that the set \(A_6\) is the set of all elements of \(B\) whose product with at least one element of \(D\) is between \(50\) and \(100\text{,}\) exclusive. So, to find the elements of \(A_6\text{,}\) we need to calculate the products of elements of \(B\) with elements of \(D\text{.}\)
    Once we calculate that we see, \(A_6=\set{7,11,13,17,19}\text{.}\)
  7. \(A_7\) is the set of all integers \(n\) satisfying the inequality \(n^2-5n-16\leq n\text{,}\) or in other words, \(n^2-6n-16\leq 0\text{.}\) If we factor the expression on the left, we see that we want to find the integers satisfying \((n+2)(n-8)\leq 0\text{.}\) Then, we see that for this inequality to be valid, we need to have \(n\in[-2,8]\text{.}\)
    Therefore, \(A_7=\set{-2,-1,0,1,2,3,4,5,6,7,8 }\text{.}\)

1.4.2.

Scratchwork.
  1. We see that this set is the set of positive multiples of \(5\text{,}\) or the set of all natural numbers that are multiples of \(5\text{.}\) Therefore we can write this set \(A=\set{5n\st n\in\mathbb{N}}\text{,}\) or \(A=\set{n\in\mathbb N\st n=5m \text{ for some } m\in\mathbb{N}}\text{.}\)
    We can also use different representations for this set (for any set really), but keeping our definition for the set accurate and concise makes it easier to understand and use.
  2. We see that \(B\) is the set of all natural numbers from \(10\) to \(100\text{,}\) inclusive.
    Therefore we can write this set as \(B=\set{n\in\mathbb N\st 10\leq n\leq 100}\) or, equivalently, \(B=\set{n\in\mathbb Z\st 10\leq n\leq 100}\text{.}\)
  3. Even though it may seem like there is not a connection between the elements of this set. We see that the elements are all \(1\) away from the square of a natural number.
    Therefore we can write this set as \(C=\set{n\in\mathbb N \st n=m^2-1 \text{ for some } m\in\mathbb N}\text{,}\) or, equivalently, \(C=\set{n^2-1\st n\in\mathbb N}\text{.}\)
  4. We see that the elements of this set are fractions. So, to understand how we can write this set in set builder notation, we can try to find a connection between the numerator and the denominator of the elements of the set. Indeed, we see that \(5=2^2+1\text{,}\) \(10=3^2+1\text{,}\) \(17=4^2+1\text{,}\) etc..
    Therefore, we can write the set as \(D=\set{\frac{m}{m^2+1} \st m \in \mathbb Z}\text{.}\)
  5. In this set we see that we only have natural numbers. We also see that all the elements of the set are powers of \(2\) (even though the fact \(65535=2^{16}\) may not be common knowledge, it is easy to check once we the `hunch’ it may be). But, we also see that not all the powers of \(2\) are in the set. For example \(32=2^5\) is not in the set. So, if we write the set as powers of \(2\text{,}\) we see \(E=\set{2^1, 2^2, 2^4, 2^8, 2^{16}, \ldots}\text{.}\) This suggests that the powers themselves are also powers of \(2\text{.}\)
    Thus, we can write this set as \(E=\set{n\in\mathbb{N}\st n=2^{(2^k)} \text{ for some nonnegative integer } k}\text{.}\)
  6. In this example, we see that every element of the set is a natural number and also divisible by \(2\) or \(3\text{.}\) This may suggest we write the set as \(F=\set{n\in\mathbb N\st n=2k \text{ or } n=3k \text{ for some } k\in\mathbb N}\text{.}\) However, we see that this is not the right definition since we know that \(2\) divides \(10\text{,}\) but \(10\) is not in the set. So we realise that this is the set of all natural numbers that are only divisible by \(2\) or \(3\text{.}\)
    Therefore we can write this set as \(F=\set{2^a 3^b \st (a,b \in \mathbb Z), (a,b \geq 0), \text{ and }(a+b\neq 0)}\text{.}\) Here, we are taking \(a+b\neq 0\) so that \(a,b\) are not both \(0\text{,}\) since \(1\notin F\text{.}\)

2 A little logic
2.7 Exercises

2.7.9.

Scratchwork.
  1. We cannot determine whether or not it was raining on Monday.
  2. It was not raining on Tuesday.
  3. It was raining on Wednesday.
  4. We cannot determine whether or not it was raining on Thursday.

2.7.14.

Scratchwork.
Recall that the only way an implication is false is if its conclusion is false while its hypothesis is true. So, \(P\implies (Q\land R)\) being false means that \(Q\land R\) is false while \(P\) is true. Note that \(Q\land R\) being false means that at least one of \(Q\) and \(R\) are false
Consider the second implication, \(\big((\neg Q)\land R\big) \implies (\neg P)\text{,}\) which is true. Since \(P\) is true, the conclusion of this implication, \(\neg P\) is false. But this means that the hypothesis, \((\neg Q)\land R\) is false, by modus tollens. More explicitly, if the hypothesis were true while the conclusion is false, then the implication is false, but if the hypothesis and conclusion are both false, then the implication is true. Therefore, at least one of \(\neg Q\) and \(R\) are false.
We have determined that at least one of \(Q\) and \(R\) are false, and at least one of \(\neg Q\) and \(R\) are false. In order for both of these conditions to be satisfied, \(R\) must be false. If \(R\) were true, then both \(Q\) and \(\neg Q\) would necessarily be false, which is impossible!
We cannot determine the truth value of \(Q\) from the information given. While \(P\) must be true and \(R\) must be false, \(Q\) can be true or false. We show how you can verify this:
  • Assume \(Q\) is false, then \(Q\land R\) is false and \(P\) is true, so the first implication is false. Similarly, \((\neg Q) \land R\) is false, while \(\neg P\) is false, so the second implication is true.
  • Now assume Q is true, then \(Q\land R\) is false and \(P\) is true, so the first implication is false. Similarly, \((\neg Q) \land R\) is false, while \(\neg P\) is false, so the second implication is true.
Alternatively, we could determine the answer by analyzing the truth tables of the two implications.
\(P\) \(Q\) \(R\) \(P\implies (Q\land R)\) \(\big((\neg Q)\land R\big) \implies (\neg P)\)
T T T T T
T T F F T
T F T F F
T F F F T
F T T T T
F T F T T
F F T T T
F F F T T
The second and fourth rows are the only rows where the first implication is false, and the second true. In both these rows, \(P\) is true and \(R\) is false, so the truth values of these statements are determined. Since \(Q\) is true in the second row, but false in the fourth row, we cannot determine the truth value of \(Q\text{.}\) Either \(P\) is true, \(Q\) is true, and \(R\) is false, or \(P\) is true, \(Q\) is false, and \(R\) is false.

2.7.15.

Scratchwork.
Since \(S\) is true, \(\neg S\) is false, and consequently the conclusion of the implication, \(Q\land (\neg S)\text{,}\) is false. But since the implication is true, its hypothesis \(R\lor(\neg P)\) must also be false. If the hypothesis were true while the conclusion were false, then the implication would be false.
Since \(R\lor(\neg P)\) is false, both \(R\) and \(\neg P\) are false. So \(R\) is false while \(P\) is true.
Finally, we use that \(P \iff \big(Q\lor (\neg S) \big) \) is true. Since \(P\) is true, \(Q\lor (\neg S) \) must also be true. But \(\neg S\) is false, so \(Q\) must be true.
We have deduced that \(P\) and \(Q\) are true while \(R\) is false.
Alternatively, we could solve this question by analyzing the truth tables of the implication and biconditional. We only include the rows where \(S\) is true, since this is given in the question.
\(P\) \(Q\) \(R\) \(S\) \(R\lor(\neg P)\) \(\big( R\lor(\neg P)\big) \implies \big(Q\land (\neg S)\big)\) \(P \iff \big(Q\lor (\neg S) \big) \)
T T T T T F T
T T F T F T T
T F T T T F F
T F F T F T F
F T T T T F F
F T F T T F F
F F T T T F T
F F F T T F T
The only row in which the implication and biconditional are true is row 2, so we look at this row to discern the truth values of \(P\text{,}\) \(Q\text{,}\) and \(R\text{:}\) \(P\) is true, \(Q\) is true, and \(R\) is false.

2.7.16.

Scratchwork.
We give a few different approaches.
  • If \(A\) and \(B\) are statements, recall that \(A\iff B\) is true only if both \(A\) and \(B\) are true, or both \(A\) and \(B\) are false. Applying this with \(A=(P\lor Q)\implies R\) and \(B=Q\land S\text{,}\) and using the fact that \((P\lor Q)\implies R\) is false, we deduce that \(B=Q\land S\) must be false. Now, since \(Q\land S\) is false but \(S\) is true, \(Q\) must be false.
    Finally, we’ll use the fact that \((P\lor Q)\implies R\) is false to determine the truth values of \(P\) and \(R\text{.}\) Recall that the only way that an implication is false is if its hypothesis is true, but its conclusion is false; in this case, that means the hypothesis \(P\lor Q\) is true, and the conclusion \(R\) is false. But remember that \(Q\) is false, so the only way \(P\lor Q\) is true is if \(P\) is true.
    We have figured out that \(P\) is true, \(Q\) is false, and \(R\) is false.
  • We can also start from the fact that the implication “\((P\lor Q)\implies R\)” is false. An implication is only false when its hypothesis is true but its conclusion is false. This immediately tells us that \(R\) is false. It also tells us that \(P \lor Q\) is true.
    Since the biconditional is true, but the clause on the left is false, we know that the clause on the right must also be false. That is \(Q \land S\) must be false. Since we are told that \(S\) is true, we know that \(Q \) is false. Putting this together with the fact that \(P\lor Q\) is true, we deduce that \(P\) is true.
    So again, we find that \(P\) is true, \(Q\) is false, and \(R\) is false.
  • Alternatively, we could use truth tables to determine the answer. In the following, we build the truth table of \(((P\lor Q)\implies R) \iff (Q\land S)\text{,}\) but only include the rows where \(S\) is true, since this is given in the question.
    \(P\) \(Q\) \(R\) \(S\) \((P\lor Q)\implies R\) \(Q\land S\) \(((P\lor Q)\implies R) \iff (Q\land S)\)
    T T T T T T T
    T T F T F T F
    T F T T T F F
    T F F T F F T
    F T T T T T T
    F T F T F T F
    F F T T T F F
    F F F T T F F
    The fourth row is the only row where \((P\lor Q)\implies R\) is false but \(((P\lor Q)\implies R) \iff (Q\land S)\) is true. Therefore we look to this row to determine the truth values of \(P\text{,}\) \(Q\text{,}\) and \(R\text{:}\) \(P\) is true, \(Q\) is false, and \(R\) is false.
Notice that the second solution is really just a reordering of the steps given in the first question. Since the initial deductions we made from the biconditional being true and the implication being false were independent of each other, we could make these deductions in either order. The last solution is different: rather than making logical deductions from the information given, we use the truth table to look at every possible case for the truth values of \(P\text{,}\) \(Q\text{,}\) and \(R\text{,}\) and look for the cases where the conditions given in the question are satisfied.

3 Direct proofs
3.5 Exercises

3.5.1.

Scratchwork.
We see that this is a conditional statement, and to prove it, we are going to assume that the hypothesis is true and show that the conclusion follows. This means that we assume that \(n\) is an even number and show that \(n^2+3n+5\) is odd.
This means that, we have an assumption and we need to understand what this assumption means by using the definitions and extracting useful expressions, i.e. structures we can work with.
In this question, since \(n\) is even, we know that \(n=2k\) for some \(k\in\mathbb Z\text{.}\) This is a nice structure and we can build upon it. Now, since we want to show that \(n^2+3n+5\) is odd, we can try to see what this expression is equal to given the assumption, \(n=2k\text{.}\)
We see that \(n^2+3n+5=(2k)^2+3(2k)+5=2(2k^2+3k+2)+1\text{.}\) Then, using the definition of even and odd, we see that \(n^2+3n+5\) is an odd number since \(2k^2+3k+2\in\mathbb{Z}\text{.}\) This is what we wanted to show.
Of course, we don’t need this much explanation in the actual proof. So, let’s clear this up and turn this into a nice proof.

3.5.2.

Scratchwork.
Even though this doesn’t look like a conditional statement, we wee that we can rewrite it as:
\begin{gather*} \text{if } m,n \text{ are odd, then } mn \text{ is odd.} \end{gather*}
Therefore, to prove it, we are going to assume that the hypothesis is true and show that the conclusion follows. This means that we assume that \(m,n\) are two odd numbers and show that \(mn\) is odd.
Here, we are going to rely on the definitions again to get as much information as we can from our assumption. Since \(m\) and \(n\) are odd, we know that \(m=2k+1\) for some \(k\in\mathbb Z\) and \(n=2\ell+1\) for some \(\ell\in\mathbb Z\text{.}\) Using this structure, we can get a nice expression for \(mn\text{.}\)
We see that \(mn=(2k+1)(2\ell+1)=2(2k\ell+k+\ell)+1\text{.}\) Then, using the definition of even and odd functions, we see that \(mn\) is an odd number since \(2k\ell+k+\ell\in\mathbb{Z}\text{.}\) This is what we wanted to show. Let’s clean this up.

3.5.3.

Scratchwork.
Let’s consider the first case: the sum of two odd numbers. We’re trying to determine the parity of \(m+n\text{,}\) where both \(m\) and \(n\) are odd. Since they’re both odd, we know that \(m=2k+1\) and \(n=2\ell+1\) for some integers \(k,\ell\text{.}\) Then we can write
\begin{equation*} m+n = (2k+1)+(2\ell+1)=2k+2\ell+2. \end{equation*}
We want to write \(m+n\) in either the form \(2p\) or \(2p+1\) for some integer \(p\text{,}\) which is possible depending on whether or not there is a factor of \(2\) in \(m+n\text{,}\) respectively. In this case, we can factor out a \(2\text{:}\)
\begin{equation*} m+n = 2(k+\ell+1) \end{equation*}
and so \(m+n\) is even. Note also that here we use the fact that \(k+\ell+1\) is an integer, since both \(k\) and \(\ell\) are integers.
With this scratchwork we can write a formal proof to the statement, “the sum of two odd numbers is even.” The solutions to the remaining parts can be determined using the same logic: first, we’ll suppose that \(m\) and \(n\) are two integers whose parity is given. We can then write out \(m\) in the form \(2k\) if its even, or \(2k+1\) if its odd, and similarly for \(n\) (although we should be careful not to reuse the variable \(k\) in our expression for \(n\)). Using these expressions we can then obtain an expression for \(m+n\) or \(mn\text{.}\) Then we manipulate this expression so that it’s either in the form \(2p\) or \(2p+1\text{,}\) where \(p\) is an integer, to determine if it is even or odd.

3.5.6.

Scratchwork.
This may look slightly harder then the previous questions since it involves \(5\) different variables and a conjunction (an “and”) statement. But, since this is a conditional statement, our strategy is going to be the same. We are going to assume the hypothesis to be true and show the conclusion.
So, we assume that \(n\mid a\) and \(n\mid b\text{.}\) This means that \(a=nk\text{,}\) and \(b=n\ell\) for some \(k,\ell\in\mathbb Z\text{.}\) Once we have these expressions, we can bo back and try to see what we wanted to prove again.
We see that we want to prove that \(n\mid (ax+by)\text{.}\) In other words, we want to prove \(ax+by=nm\) for some \(m\in\mathbb Z\text{.}\) Since we know that \(a=nk\text{,}\) and \(b=n\ell\text{,}\) we can write \(ax+by=nkx+n\ell y=n(kx+\ell y)\text{.}\) This is what we wanted to show. Let’s write this in a nice proof.

3.5.7.

Scratchwork.
The hypothesis tells us that \(a=nk\) and \(a+1=n\ell\text{.}\) This means we can write \(1 = (a+1)-a = n\ell - nk = n(\ell-k)\text{.}\) But this means \(n \mid 1\text{.}\) This is the basis of our proof.

3.5.8.

Scratchwork.
This is a conditional statement. To prove this, we need to assume that the hypothesis is true and show that the conclusion follows. This means that we assume \(3\mid a\) and \(2\mid a\) and show that \(6\mid a\text{.}\)
This is “one of those questions” where our previous knowledge may be misleading. We may want to say that since \(6=3\cdot 2\text{,}\) then if \(a\) is divisible by \(2\) and \(3\text{,}\) it should be divisible by \(6\text{.}\) But this would be a wrong chain of implications. Even though the result is true in this instance, it may be wrong if we change the numbers a little. For example, if we change \(3\) by \(4\text{,}\) that is, “If \(4\mid a\) and \(2\mid a\text{,}\) then \(8\mid a\)”, even though the previous chain of implications doesn’t change, the result doesn’t follow anymore (for an example you can see the statement is false for \(a=4\) ).
It is always preferable to use the definitions and the results we have proven to prove these statements. Let’s see what we can do.
Assume \(3\mid a\) and \(2\mid a\text{,}\) that is, \(a=2k\) and \(a=3m\) for some \(k,m\in\mathbb Z\text{.}\) Since we want to show that \(6\mid a\text{,}\) we can multiply both sides of the equation \(a=2k\) by \(3\) and both sides of the equation \(a=3m\) by \(2\) and get \(3a=6k\) and \(2a=6m\text{.}\) Then, subtracting the second form the first, we get \(a=3a-2a=6k-6m=6(k-m)\text{,}\) which is the desired result. Now, we can write this nicely in a proof.
Alternatively, assuming \(a=2k\) and \(a=3m\) for some \(k,m\in\mathbb Z\text{,}\) we can try to show that \(m=2\ell\) for some \(\ell\in\mathbb Z\text{.}\) We can do this by writing, \(a=2k=3m\text{,}\) and hence, \(2k-2m=2(k-m)=m\text{.}\) Therefore \(a=3m=6(k-m)\text{,}\) and the result follows.

3.5.9.

Scratchwork.
To prove this conditional statement, we are going to assume the hypothesis to be true and show that the conclusion follows. That means, we assume that \(3\mid (n-4)\) and show that \(3\mid (n^2-1)\text{.}\)
We see that assuming \(3\mid (n-4)\) means that \(n-4=3k\) for some \(k\in\mathbb Z\text{,}\) and under this assumption, we want to show that \(3\mid (n^2-1)\text{,}\) that is, \(n^2-1=3m\) for some \(m\in\mathbb Z\text{.}\) This suggests that if we have an expression on \(n\) we can simply calculate \(n^2-1\) and get the desired result (alternatively, we could try to find a connection between \((n-4)\) and \((n^2-1)\)).
Since we know that \(n-4=3k\) for some \(k\in\mathbb Z\text{,}\) we see that \(n=3k+4\text{.}\) Thus, we get \(n^2-1= (3k+4)^2-1=(9k^2+24k+16)-1= 3(3k^2+8k+5)\text{,}\) which is what we wanted to show. Of course, in a proof we don’t need this much explanation. Indeed, we can write this explanation nicely in a proof.

3.5.13.

Scratchwork.
Before we get into the proof we see that every integer \(\ell\) is an integer root since we can take \(k=1\) and \(m=\ell\) as given in the definition. But these numbers also include non-integer numbers such as \(\sqrt[3]{7}\text{,}\) since we can take \(k=3\) and \(m=7\) or \(\sqrt 5\) since we can teke \(k=2\) and \(m=5\) as given in the definition. We also see that the statement also says that the product of \(\sqrt[3]{7}\) and \(\sqrt 5\) should also be an integer root. If we check carefully, we see that
\begin{equation*} \sqrt[3]7\cdot \sqrt 5=7^{1/3}5^{1/2}=(7^2)^{1/6}(5^3)^{1/6}=(7^25^3)^{1/6}. \end{equation*}
Indeed, the product is an integer root.
Now, assume that \(a\) and \(b\) are integer roots. This means that there are natural numbers \(k,\ell\) and integers \(m,n\) such that \(a^k=n\) and \(b^\ell=m\text{.}\) Now, we are going to do the same trick we did in the example above to make the powers “same” so that we can multiply the bases. We see \(a^{k\ell}=n^\ell\) and \(b^{\ell k}=m^k\text{.}\) Thus, \((ab)^{k\ell}=n^\ell m^k\text{,}\) which is the desired result, since \(n^\ell m^k\in\mathbb Z\text{.}\) Now, we can write this in a proof.

3.5.16.

Scratchwork.
It is a common (and easy to make) mistake to apply any operation on an inequality and implicitly assuming that the operation keeps the order of the inequality. Namely, if \(f(x)\) is a function and we know \(x\lt y\text{,}\) we may want to say \(f(x)\lt f(y)\text{.}\) Unfortunately this is not always true, and we have to be careful. This reasoning would only be true if the function \(f\) is an increasing function in the given domain; and this question tells us that \(f(x)=\sqrt{x}\) is, indeed, an increasing function.
We see that what we want to prove in this question is a conditional statement. So, we can assume the hypothesis and try to show the conclusion. So, we assume \(x\gt y\) and try to show that \(\sqrt{x} \gt \sqrt{y}\text{.}\) Since we cannot simply take the square roots of both sides (because of the above argument), we need to find a different way to prove the conclusion.
We see that we can rewrite the hypothesis as \(x-y\gt 0\text{.}\) We also know that
\begin{equation*} \underbrace{x-y}_{\gt 0}=(\sqrt{x})^2-\sqrt{y})^2=(\sqrt{x}-\sqrt{y})\underbrace{(\sqrt{x}+\sqrt{y})}_{\gt 0}, \end{equation*}
where we have factored the difference of squares: \(a^2-b^2 = (a-b)(a+b) \text{.}\) This tells us that \(\sqrt{x}-\sqrt{y}\gt 0\text{,}\) which is what we wanted to show. Now, we can write this nicely in a proof.

3.5.18.

Scratchwork.
For scratchwork, we’ll begin with the inequality
\begin{equation*} \sqrt{x+y}\leq \sqrt{x}+\sqrt{y} \end{equation*}
Note that this is what we actually want to show though! When we write up the formal proof, we’ll have to reverse the order of our logic. We’ll try to derive something true from this inequality, which will be the starting point of our actual proof. Starting with
\begin{equation*} \sqrt{x+y}\leq \sqrt{x}+\sqrt{y}, \end{equation*}
note that \(0\leq \sqrt{x+y}\text{,}\) and so we can square both sides to obtain
\begin{equation*} x+y \leq x+2\sqrt{x}\sqrt{y}+y. \end{equation*}
Here we are using that \((\sqrt{x})^2=x\) since \(x\geq0\text{,}\) and similarly for \(y\text{.}\) Subtracting \(x+y\) from both sides of this inequality, we obtain
\begin{equation*} 0 \leq \sqrt{x}\sqrt{y} \end{equation*}
and this inequality is true since the square root function is never negative. So, for the actual proof of the statement, we should start with the inequality
\begin{equation*} 0\leq\sqrt{x}\sqrt{y}, \end{equation*}
and reverse the argument above to show the desired conclusion,
\begin{equation*} \sqrt{x+y}\leq \sqrt{x}+\sqrt{y}. \end{equation*}

5 More proofs
5.5 Exercises

5.5.1.

Scratchwork.
We see that this is a conditional statement and as we did in the previous chapters, we can try direct proof where we assume the hypothesis and show the conclusion. That means, we assume that \(n^2+4n+5\) is odd and show that \(n\) is even.
This means that we assume \(n^2+4n+5=2k+1\) for some \(k\in\mathbb Z\) and show that \(n=2m\) for some \(m\in\mathbb Z\text{.}\) However, to get from \(n^2+4n+5=2k+1\) to showing that \(n\) is even will not an easy task. It would require us to take a square root, and so we would also need to understand the effect of taking square root on the parity of the number. Oof! That would be a whole other exercise.
So, we may need to look at a different proof method. Since this is a conditional statement, we can look at its contrapositive and see whether it simplifies the problem. First, let \(n\in\mathbb Z\text{,}\) then the contrapositive is:
\begin{gather*} n \text{ is not odd } \implies n^2+4n+5 \text{ is not odd}. \end{gather*}
But since \(n\) is an integer, we know that \(n^2+4n+5\) is an integer. So if \(n\) and \(n^2+4n+5\) are not odd, then they must be even. Hence we can write the contrapositive as
\begin{gather*} n \text{ is odd } \implies n^2+4n+5 \text{ is even}. \end{gather*}
This immediately looks like a simpler statement since the hypothesis in on \(n\text{,}\) which is the simpler term, and the conclusion is on \(n^2+4n+5\text{,}\) which is the more complex term. From here, the rest of the proof is straightforward. We assume \(n=2k+1\) for some \(k\in\mathbb Z\) and then we write the expression for \(n^2+4n+5\) and show that it is even. Namely,

5.5.2.

Scratchwork.
Trying to prove this statement directly doesn’t make sense, and instead we’re going to prove this statement by its contrapositive; that way, we will end up with a statement about what happens when \(5\) does divide an integer, and we can directly use the definition of that.
The contrapositive of this statement is: If \(5\mid n\text{,}\) then \(5\mid n^2\text{.}\) We’ve had plenty of practice proving statements like this.

5.5.3.

Scratchwork.
In this question we see that we have a conditional statement. Hence, we can assume the hypothesis and try to show the conclusion. In other words, we can assume \(5\nmid n\) or \(2\nmid n\) and try to show that \(10\nmid n\text{.}\) But, we we have seen before, whenever we have a hypothesis that is an “or”-statement, we may need to look at cases. Moreover, since each part of the hypothesis is of the form \(a\nmid b\text{,}\) we may need to look at cases for those as well. This means that if we want to prove this statement using direct proof and cases, it may get very long. Also, when we have lots of cases, the chances of making a mistake go up.
However, we can also look at the contrapositive of the statement and see whether it may be easier to prove. For \(n\in\mathbb Z\text{,}\) the contrapositive of the statement of the statement is:
\begin{gather*} \text{if } 10\mid n, \text{ then } 5\mid n \text{ and } 2 \mid n, \end{gather*}
which looks much more promising, since if a number is divisible by \(10\text{,}\) then it should be divisible by \(5\) and \(2\text{.}\) Let’s see how we can put that in a proof.

5.5.4.

Scratchwork.
We see that this is a conditional statement. First let \(n,m\in\mathbb N\text{.}\) So, to prove the statement, we can assume the hypothesis, i.e. \(n \neq 1\) and \(n \neq 2\) and show that \(n\nmid m\) or \(n \nmid (m+2)\text{.}\) However, the assumption doesn’t give us much to work with. We can try to look at cases to create more structure to work with, but unfortunately, it is also not clear what those cases should be.
Changing our strategy, we can also look at the contrapositive of the statement:
\begin{equation*} \text{If } n\mid m \text{ and } n\mid (m+2), \text{ then } n=1 \text{ or } n=2. \end{equation*}
Here we see that we assume \(n\mid m\) and \(n\mid (m+2)\text{.}\) Then we can write \(m=xn\) and \(m+2=yn\text{.}\) Now since the result says something about \(n,1\) and \(2\text{,}\) it makes sense to try to use this equations to say something about \(2\text{.}\) Taking the difference of these gives us
\begin{equation*} 2 = xn-yn = n(x-y) \end{equation*}
which implies that \(n\) should be able to divide \(2\) as well. But we know only natural number divisors of \(2\) are \(1\) and \(2\text{,}\) which is just what we wanted to show. Now, let’s write this up.

5.5.5.

Scratchwork.
We see that this is a conditional statement. To prove it, we can assume that \(n^2+m^2\) is even and try to show that then \(n,m\) have the same parity, that is, \(n,m\) are both odd or both even. We know that \(n^2+m^2\) being even means that \(n^2+m^2=2k\) for some \(k\in\mathbb Z\text{.}\) We see that it may not be trivial to get from this assumption to the conclusion on \(n,m\text{.}\) It would require us to take square-roots or solve a quadratic or something like that. Urgh.
What we may try instead, is to look at the contrapositive of the statement we want to prove:
\begin{gather*} \text{if } n,m \text{ have different parities, then } n^2+m^2 \text{ is not even,} \end{gather*}
and since we know that \(n^2+m^2\in\mathbb Z\text{,}\) the conclusion also means that \(n^2+m^2\) is odd.
Here, if we assume the hypothesis of the contrapositive, we see that we have \(2\) cases: \(n\) is even and \(m\) is odd, and \(m\) is even and \(n\) is odd.
  • If \(n\) is even and \(m\) is odd, then we have \(n=2a\) and \(m=2b+1\) for some \(a,b\in\mathbb Z\text{.}\) This means that \(n+2+m^2=(2a)^2+(2b+1)^2=4a^2+4b^2+4b+1=2(2a^2+2b^2+2b)+1\text{,}\) which is odd.
  • If \(n\) is odd and \(m\) is even, then we have \(n=2a+1\) and \(m=2b\) for some \(a,b\in\mathbb Z\text{.}\) This means that \(n+2+m^2=(2a+1)^2+(2b)^2=4a^2+4a+1+4b^2=2(2a^2+2b^2+2a)+1\text{,}\) which is also odd.
Here, we also see that these cases are almost the same since the conclusion is symmetric with respect to \(n,m\text{.}\) So, practically, we will only need to prove one of case, and the other will be proven very similarly. So, we can use “WLOG” and prove only one of the cases. But, whenever you want to use “WLOG”, you should always convince yourself that the cases are similar by doing both possibilities and seeing just how similar they are. Let’s see how we can make this work in a proof.

5.5.6.

Scratchwork.
If we were to provide a direct proof of this statement, we’d have to start with the inequality \(x^3+5x\geq x^2+1\) which would be difficult, since we have a cubic term. Instead we’re going to prove this statement by its contrapositive; this will allow us to assume that \(x\leq 0\text{,}\) and from that we will know the sign of powers of \(x\text{.}\)
The contrapositive of this statement is: If \(x\leq 0\text{,}\) then \(x^3+5x \lt x^2+1\text{.}\) Notice that the right-hand side is positive, since \(x^2 \geq 0\text{.}\) At the same time, since \(x \leq 0\text{,}\) we know that \(x^3 \leq 0\) and \(5x \leq 0\text{.}\) So the left-hand side is non-positive. Thus we can write \(x^3+5x \leq 0 \lt x^2+1\text{.}\) And now we have everything we need for the proof.

5.5.7.

Scratchwork.
Proving things directly looks difficult, so we look at the contrapositive: If \(a\) and \(b\) are consecutive, then \(a+b\) is odd. Now we can use the definition of consecutive to prove things.

5.5.8.

Scratchwork.
We see that this is a conditional statement. This means that we can assume the hypothesis and then we try to show the conclusion. Hence, we assume that \(n=2k\) for some \(k\in\mathbb Z\text{,}\) and show that \(n=4m\) or \(n=4m+2\) for some \(m\in\mathbb Z\text{.}\) However, when we wrote \(n=2k\) for some \(k\in\mathbb Z\text{,}\) we have extracted all the information from our assumption and we still need more to finish the proof.
Proof by cases is a very useful tool in situations like these since it creates extra structure for us to work with.
In this question, if we pay a little closer attention to the conclusion of the statement, we see that we want to show is \(n=4m=2(2m)\) or \(n=2(2m+1)\) for some \(m\in\mathbb Z\text{.}\) Moreover, since \(k\in\mathbb Z\text{,}\) we know that \(k=2m\) or \(k=2m+1\) for some \(m\in\mathbb Z\text{,}\) which suggests that we can use cases to finish the proof.

5.5.9.

Scratchwork.
Since this is a biconditional statement, we need to prove the implication in both direction. That is we must prove
\begin{align*} \big(2\mid (n^4-7)\big) \amp \implies \big(4\mid (n^2+3)\big). \amp \text{and} \amp \amp \big(4\mid (n^2+3)\big) \amp \implies \big(2\mid (n^4-7)\big) \end{align*}
We’ll prove each implication in turn.

5.5.10.

Scratchwork.
We see that this is a biconditional statement and thus we need to prove both implications in both direction. That is, we need to prove
\begin{gather*} (3\mid 5a)\implies(3\mid a) \qquad \text{ and } \qquad (3\mid a)\implies (3\mid 5a). \end{gather*}
We can see that the second statement should be quite straightforward, as if a number is a multiple of \(3\) then \(5\) times that number will also be a multiple of \(3\text{.}\)
However, the first statement may be a little more involved. We can try to look at the contrapositive of the statement:
\begin{gather*} (3\nmid a)\implies(3\nmid 5a). \end{gather*}
So that we have an assumption on the simple term \(a\text{.}\) But, this also suggests that we use cases: \(a=3k+1\) or \(a=3k+2\) for some \(k\in\mathbb Z\text{.}\) Then, we can calculate \(5a\) and show that it is also not divisible by \(3\text{.}\)
We can also try to prove it directly by assuming \(3\mid 5a\text{,}\) that is, \(5a=3k\) for some \(k\in\mathbb Z\) and then manipulating this equation to show \(a=3n\) for some \(n\in\mathbb Z\text{.}\) Starting from \(5a=3k\text{,}\) add \(a\) to both sides to get \(6a=3k+a\text{.}\) Now subtract \(3k\) from both sides to get \(6a-3k=a\text{.}\) That is \(a=3(2a-k)\text{.}\) Since \(2a-k \in \mathbb Z\text{,}\) we have shown that \(3 \mid a\text{.}\) Let’s put everything together.

5.5.11.

Scratchwork.
In order to show that \((n^2-1)(n^2+2n)\) is divisible by \(4\text{,}\) we need to show that \((n^2-1)(n^2+2n)=4k\) for some \(k\in\mathbb{Z}\text{.}\) This looks complicated, so we should split it into cases!
Factoring the expression, we have
\begin{equation*} (n^2-1)(n^2+2n)=(n-1)(n+1)n(n+2) \end{equation*}
This is a product of four consecutive terms! Using this factorization, we could do the proof by looking at four cases \(n=4k\text{,}\) \(4k+1\text{,}\) \(4k+2\) and \(4k+3\text{.}\) That will work, but can we complete the proof in fewer cases?
Since the polynomial is the product of two factors, it would be sufficient to show that both are even, or that one of them is divisible by 4. This suggests that we might get away with considering the parity of \(n\) — what happens when \(n\) is even, and when \(n\) is odd. Lets try that first.
  • In order to show that \((n^2-1)(n^2+2n)\) is divisible by \(4\text{,}\) we need only show one of these factors is divisible by \(4\text{.}\)
  • When \(n\) is even \(n^2\) is even, so \(n^2-1\) is odd, and not divisible by \(4\text{.}\) We will need to show that \(n^2+2n\) is divisible by \(4\) in this case. That isn’t too hard to show.
  • When \(n\) is odd \(n^2\) is odd, and so \(n^2+2n\) is odd, and not divisible by \(4\text{.}\) We will need to show that \(n^2-1\) is divisible by \(4\) in this case. Since \(n\) is odd, we know \(n=2k+1\text{,}\) so \(n^2-1 = 4k^2+4k+1-1\text{,}\) so this isn’t too hard.

5.5.12.

Scratchwork.
We would like to prove the contrapositive, but first we need to figure out what the contrapositive statement is. Following the technical definition, we know that the contrapositive is If \(\neg\)(\(x\) or \(y\) is odd, but not both), then \(\neg\)(\(x+y\) is odd). But we would like to translate this into a statement that is easier to comprehend.
There are four possibilities for the parities of \(x\) and \(y\text{:}\)
  1. \(x\) and \(y\) are both even.
  2. \(x\) is even and \(y\) is odd.
  3. \(x\) is odd and \(y\) is even.
  4. \(x\) and \(y\) are both odd.
Options 2 and 3 fit the condition \(x\) or \(y\) is odd, but not both. Therefore, \(\neg\)(\(x\) or \(y\) is odd, but not both) is satisfied by conditions 1 and 4. We can summarize these two conditions as “\(x\) and \(y\) have the same parity.” Moreover, we can rewrite the statement \(\neg\)(\(x+y\) is odd) as “\(x+y\) is even.”
Putting these together, we get the contrapositive: If \(x\) and \(y\) have the same parity, then \(x+y\) is even.

5.5.13.

Scratchwork.
We see that this is a conditional statement. So, to prove that we can assume the hypothesis and show the conclusion. That is, we assume that \(3\mid (n^2+4n+1)\) and try to show that \(n\equiv 1\mod 3\text{.}\) We see that our assumption means that \(n^2+4n+1=3m\) for some \(m\in\mathbb Z\text{.}\) From this assumption, we need to get to our conclusion on \(n\text{.}\) Even though it is not impossible, it will require factoring or taking the square root, and we may need to prove more statements about how those operations affect divisibility by \(3\text{.}\)
Instead, we can look at the contrapositive of the statement and try to have an assumption on the simpler term \(n\text{.}\) First let \(n\in\mathbb Z\text{.}\) Then, we see that the contrapositive of the statement is
\begin{gather*} \text{if } n\not\equiv 1\mod 3\text{ then } 3\nmid (n^2+4n+1). \end{gather*}
Therefore we assume that \(n\not\equiv 1\mod 3\text{.}\) This means that we have \(2\) cases, \(n \equiv 0\mod 3\) or \(n \equiv 2\mod 3\text{.}\) We see that both of these cases give us nice structure to work on and say something about \(n^2+4n+1\text{.}\) Let’s work this out in a proof.

5.5.14.

Scratchwork.
We see that this is a conditional statement. As we did before we can try to prove it directly by assuming the hypothesis and proving the conclusion. This means that we assume \(5\nmid m\) and show that \(m^2\equiv 1 \mod{5}\) or \(m^2\equiv -1 \mod{5}\text{.}\)
Since the hypothesis is \(5\nmid m\text{,}\) this suggests we may need to use cases to prove the result. Once we divide the proof into cases, the rest becomes quite straightforward.
We could also think about looking at the contrapositive of this statement. This may make sense since the conclusion of the statement is an “or”-statement. We see that the contrapositive of the statement is:
\begin{gather*} \text{if } m^2\not\equiv 1 \mod{5} \text{ and } m^2\not\equiv -1 \mod{5}, \text{ then }5\mid m. \end{gather*}
This still suggests that we use case, and, on top of that, we would have assumptions on the more complex term, \(m^2\text{,}\) instead of the simpler term, \(m\text{.}\) Therefore, using direct proof with cases may be easier in this question. Let’s write the proof using cases.

5.5.15.

Scratchwork.
Before attempting the question, we should think about what method of proof will be easiest to show. Based on the content of the chapter, it might be wise to consider a proof by contrapositive or a direct proof with cases.
The contrapositive statement is: If \(q^2 \not \equiv 1 \mod 3\text{,}\) then \(3\mid q\). This means that either \(q^2 \equiv 0 \mod 3\) or \(q^2 \equiv 2 \mod 3\text{.}\) Working through this problem would be difficult because we start with cases on \(q^2\) and we need to get information about \(q\) from there. For this reason, we look into another method of proof.
Now, we consider what a direct proof with cases might look like. If \(3 \nmid q\text{,}\) then when we divide \(q\) by \(3\text{,}\) we will have a remainder of \(1\) or \(2\text{.}\) This gives two very clear cases. It is also easier in general to start with information about \(q\) and work towards information about \(q^2\text{.}\) Because of this, we proceed directly using cases.

5.5.18.

Scratchwork.
Questions of this type are a good examples of how useful the triangle inequality can be. We see that if we wanted to use triangle inequality we could just say:
\begin{gather*} |x+4|+|x-3|=|x+4|+|3-x|\geq |(x+4)+(3-x)|=|7| = 7. \end{gather*}
However, we are not allowed to use triangle inequality in this question. This means that we need to find a way to simplify the absolute values in the question. For that, we need to understand when the expressions \((x+4)\) and \((x-3)\) change signs. We see that these expressions change signs at \(x=-4\) and \(x=3\text{,}\) respectively.
It is quite common to think here to look at cases \(x\lt -4\text{,}\) \(x\geq 4\text{,}\) \(x\lt 3\text{,}\) and \(x\geq 3\text{.}\) However, this is not the most efficient way to divide this hypothesis into cases since even when we look at \(x\geq -4\text{,}\) we need to consider the sign of \(x-3\text{.}\) This means that within some of the cases, we may have extra cases to consider.
Instead, we can divide these into \(3\) cases: \(x\lt -4\text{,}\) \(-4\leq x\leq 3\text{,}\) and \(x\gt 3\text{.}\) This is more efficient since the expressions \((x+4)\) and \((x-3)\) do not change signs in the intervals \((-\infty,-4)\text{,}\) \([-4, 3]\text{,}\) and \((3,\infty)\text{.}\)
Let’s see how this works in the proof.

5.5.19.

Scratchwork.
We assume the hypothesis is true and then try to reach the conclusion. So we assume \(|x-1| \lt 1\) and then we have to bound \(|x^2-1|\text{.}\) In order to bound \(|x^2-1|\text{,}\) we can factor the expression, and try to bound each term.
\begin{equation*} |x^2-1|=|(x-1)(x+1)|=|x-1|\cdot|x+1| \end{equation*}
The factor \(|x-1| \lt 1\text{,}\) by assumption. But how can we bound the factor \(|x+1|\text{?}\) We’ll need to use the inequality \(|x-1| \lt 1\) again.
Recall that \(|x-1| \lt 1 \) is equivalent to the inequality
\begin{equation*} -1 \lt x-1 \lt 1. \end{equation*}
Adding \(2\) to everything, we end up with
\begin{equation*} 1 \lt x+1 \lt 3. \end{equation*}
In particular, this implies that \(|x+1| \lt 3\text{.}\)
There is a slicker way to obtain this inequality: we’ll use a common trick, where we “add zero in a fancy way,” in order to introduce the term \(x-1\text{.}\) Then, we’ll use the triangle inequality and the bound \(|x-1| \lt 1\text{.}\)
\begin{equation*} |x+1| = |(x-1)+2| \leq |x-1|+|2| \lt 1+2=3. \end{equation*}
With this scrap work, we’re ready to write up the proof of the statement. Remember to be careful with the logic of the proof! We need to start by assuming the hypothesis, \(|x-1| \lt 1\text{,}\) and show the conclusion is true, that \(|x^2-1| \lt 3\text{.}\)

5.5.20.

Scratchwork.
In order to bound \(|2x^2-3x-2|\text{,}\) we can factor the expression, and try to bound each term.
\begin{equation*} |2x^2-3x-2|=|(2x+1)(x-2)|=|2x+1|\cdot|x-2| \end{equation*}
The factor \(|x-2| \lt 1\text{,}\) by assumption. But how can we bound the factor \(|2x+1|\text{?}\) We’ll need to use the inequality \(|x-2| \lt 1\) again.
Recall that \(|x-2| \lt 1 \) is equivalent to the inequality
\begin{equation*} -1 \lt x-2 \lt 1. \end{equation*}
Adding \(2\) to everything, we end up with
\begin{equation*} 1 \lt x \lt 3 \end{equation*}
which implies that \(|x| \lt 3\text{.}\) Now we can use the triangle inequality to bound \(|2x+1|\text{:}\)
\begin{equation*} |2x+1|\leq |2x|+1=2|x|+1 \lt 2\cdot3+1=7. \end{equation*}
With this scrap work, we’re ready to write up the proof of the statement. But remember, we have to make sure that our logic flows in the correct direction. We must make sure our proof starts from the hypothesis, \(|x-2| \lt 1\text{,}\) and ends at the conclusion, \(|2x^2-3x-2| \lt 7\text{.}\)

5.5.21.

Scratchwork.
We start with some scratchwork. We are asked to prove an inequality involving absolute values. Often, the triangle inequality can be helpful in these types of situations. Recall that the triangle inequality states: For any \(a,b\in \mathbb{R}\text{,}\) \(|a+b|\leq |a|+|b|\text{.}\) Following the hint, we would like to show that \(-|x-y|\leq |x|-|y|\) and \(|x|-|y|\leq |x-y|\text{.}\)
We rearrange the triangle inequality and attempt to get the formula \(|x|-|y|\leq |x-y|\) (the other formula can be obtained similarly).
First, we want to rearrange the inequality to have a minus sign on the left, and only one set of absolute values on the right:
\begin{equation*} |a+b| - |b| \leq |a|. \end{equation*}
Then we can try to write \(a\) and \(b\) in terms of \(x\) and \(y\) to make the equation above look more like the equation we want to prove. Notice that we currently have \(|a|\) on the right hand side, whereas the expression we want to obtain has \(|x-y|\text{.}\) Let’s try setting \(a = x-y\text{:}\)
\begin{equation*} |x-y+b| - |b| \leq |x-y|. \end{equation*}
Now, the left hand side has \(|x-y+b|-|b|\) where we want \(|x|-|y|\text{.}\) We set \(b = y\) to obtain the desired expression:
\begin{equation*} |x-y+y| - |y| = |x|-|y|\leq |x-y|. \end{equation*}
All that’s left to do is to write this up as a proof!

5.5.22.

Scratchwork.
The statement as written is false, which we can see by looking at the value of the function for a negative point in the domain, say \(x=-1,\) and then at a positive point in the domain, say \(y=1\text{.}\) Then \(x\leq y\) but \(f(-1)=-1\lt 1=f(1)\text{.}\) So \(f\) is not decreasing on all of \(\mathbb{R} - \{0\}\text{.}\) We will not run into this issue if we instead restrict the domain of the function to \((0,\infty)\text{.}\)
Let’s rewrite the statement as this: Let \(f:(0,\infty)\to\mathbb{R}\) be defined by \(f(x)=1/x\text{.}\) Then \(f\) is decreasing.
We are going to take \(0\lt x\lt y\) and we want to show that \(1/x\gt 1/y\text{.}\) To do this start from \(x\lt y\) and then divide both sides by \(xy\) (which is strictly positive). This gives \(1/y \lt 1/x\) as required. Notice that this argument fails precisely when \(xy\) is negative, and that happens when \(x\) and \(y\) have opposite signs - exactly as we saw in our counter-example above.

6 Quantifiers
6.6 Exercises

6.6.1.

Scratchwork.
We see that this statement is a universal statement. Since we don’t have any information on \(n\) other than being an integer, we see that we don’t really have a useful assumption on \(n\text{.}\) For example, if \(n\) were even, then we could say \(n=2m\) for some \(m\in\mathbb Z\text{.}\) But in this question, \(n\) is just an integer, which is a broad assumption without much structure.
One strategy to create such structure is to divide our assumption into cases. This may create extra steps, but eventually may help us have “something” that we can work with. In this example, since we are trying to prove a statement on divisibility by \(3\text{,}\) it would make sense to look at cases that makes use of division algorithm and divisibility by \(3\text{.}\)

6.6.4.

Scratchwork.
Even though this result looks counter-intuitive, it is indeed true.
We are going to use proof by contrapositive. We see that the expression \(\forall a,b\in\mathbb Z\) is not a part of the conditional statement, and therefore we see that the contrapositive of the statement is:
\(\forall a,b\in\mathbb Z\text{,}\) if \(3\nmid a\) or \(3\nmid b\text{,}\) then \(\forall a,b\in\mathbb Z\text{,}\) we have \(3\nmid (a^2+b^2)\text{.}\)
Looking at the contrapositive still doesn’t sound like the right thing to do since we are trading a “\(3\) divides (blah)” type of statement as the hypothesis of the conditional statement, with two “\(3\) doesn’t divide (blah)” type of statement that are connected with “or”. This may seem like it will make our assumption more complicated. But, by looking at the contrapositive, we now have an assumption on simpler building blocks, \(a\) and \(b\text{,}\) instead of the their complicated, complex, counterpart \(a^2+b^2\text{.}\) Since we now have “\(3\nmid a\) or \(3\nmid b\)”, it also suggests that we look at cases.
Since these two cases, \(3\nmid a\) and \(3\nmid b\text{,}\) are very similar we will will make the proof a little shorter by using “without loss of generality”. To make it shorter still, we will contract “without loss of generality” to “WLOG”. Note that one should always be careful using WLOG since it is easy to introduce errors by assuming two cases are very similar when, in fact, there are subtle differences between them.
Once we assume, WLOG that \(3\nmid a\text{,}\) this will introduce \(2\) new cases. Then, within those cases, we may still need to have some extra cases for the choices of \(b\text{.}\) We want to mention that this is quite standard to have cases within cases, within cases, within cases, within…

6.6.5.

Scratchwork.
The converse of the statement is:
For any \(n,a,b\in\mathbb{Z}\text{,}\) if \(n\mid ab\text{,}\) then \(n\mid a\) or \(n\mid b\text{.}\)
While the negation of this statement is:
There are some \(n,a,b\in\mathbb{Z}\) such that \(n\mid ab\) but \(n\nmid a\) and \(n\nmid b\text{.}\)
We can test some values of \(n\text{,}\) \(a\text{,}\) and \(b\) to see if the converse or its negation is true. Suppose \(n=4\text{.}\) We can factor \(n=2\cdot 2\text{.}\) This means is we take \(a=b=2\text{,}\) then \(n=ab\) and so \(n\mid ab\text{.}\) But \(4\nmid 2\text{,}\) so \(n\nmid a\) and \(n\nmid b\text{.}\) Therefore we see that the converse of the original statement is false. Now we write up a formal disproof.

6.6.6.

Scratchwork.
We’ll try to construct a pair of function \(f,g\) so that \(f+g\) is odd, but one of them, say \(f\text{,}\) is even. We can test the result by look at some simple odd and even functions. An example of an odd function is \(y=x\text{,}\) and an example of an even function is \(y=x^2\text{.}\) Let’s suppose that \(f(x)=x^2\text{,}\) which is even. We want to choose \(g\) so that \(f+g=x\text{,}\) an odd function. Then we can take \(g(x)=x-f(x)=x-x^2\text{.}\) Also, we can show that \(g\) is neither even nor odd; \(g(2)=-2\) but \(g(-2)=-6\text{.}\)

6.6.16.

Scratchwork.
Suppose that \(\epsilon \gt 0\) is given. We need to find some \(\delta \gt 0\) so that
\begin{equation*} 0 \lt |x-4| \lt \delta \implies |-3x+5-(-7)| \lt \epsilon. \end{equation*}
We can simplify the inequality on the right-hand side as
\begin{equation*} |-3x+12| \lt \epsilon. \end{equation*}
Moreover, we can factor out \(|-3|\) from the absolute value, and we’ll end up with a factor of \(|x-4|\text{:}\)
\begin{equation*} 3|x-4| \lt \epsilon. \end{equation*}
So in order to satisfy the initial implication, we need to find \(\delta \gt 0\) so that
\begin{equation*} 0 \lt |x-4| \lt \delta\implies 3|x-4| \lt \epsilon. \end{equation*}
From this, we see that this implication will be satisfied when \(\delta \leq \epsilon/3\text{.}\) So let’s just take \(\delta = \epsilon/3\text{.}\)
Now we can write up the formal proof.

6.6.17.

Scratchwork.
We need to show the following:
For any \(\epsilon \gt 0\text{,}\) there is some \(\delta \gt 0\) so that whenever \(0 \lt |x-1| \lt \delta\text{,}\) we have \(|x^2-1| \lt \epsilon\text{.}\)
In order to bound \(|x^2-1|\text{,}\) we can factor the expression, and try to bound each term.
\begin{equation*} |x^2-1|=|(x-1)(x+1)|=|x-1|\cdot|x+1| \end{equation*}
The factor \(|x-1| \lt \delta\text{,}\) by assumption. But how can we bound the factor \(|x+1|\text{?}\) The idea is this: as long as \(x\) is close to \(1\text{,}\) say \(|x-1|\leq 1\text{,}\) then \(|x+1|\) can’t be too large. When we decide how to choose delta, we don’t have to make our choice immediately; we can impose several conditions on delta, and then when we have worked out all those conditions, we can make our final choice. So, we are going to require that \(\delta\leq1\text{,}\) to ensure that \(|1+x|\) is not too large. To see this assume that \(|x-1|\leq1 \text{.}\) Then
\begin{equation*} -1\leq x-1 \leq1. \end{equation*}
Adding \(2\) to everything, we end up with
\begin{equation*} 1\leq x+1 \leq 3. \end{equation*}
In particular, this implies that \(|x+1|\leq 3\text{.}\)
So let’s assume that \(|x-1| \lt \delta\) and \(|x-1|\leq 1\text{,}\) the latter implying that \(|x+1|\leq3\text{.}\) Returning back to the factorization of \(|x^2-1|\text{,}\) we have
\begin{equation*} |x^2-1|=|x+1|\cdot|x-1| \lt 3\delta \end{equation*}
So we want to choose delta so that
\begin{equation*} 3\delta \leq \epsilon \end{equation*}
and so we are going to place another condition on delta, namely that
\begin{equation*} \delta \leq \frac{\epsilon}{3}. \end{equation*}
Now we don’t immediately set \(\delta=\epsilon/3\text{,}\) since we also required (above) that \(\delta\leq1\text{.}\) We can satisfy both of these requirements by setting \(\delta=\min\{1,\epsilon/3\}\text{.}\)
With this scratchwork, we’re ready to write up the proof.

6.6.18.

Scratchwork.
Given \(\epsilon \gt 0\text{,}\) we need to find a number \(N\) so that whenever \(n\gt N\text{,}\) we have
\begin{equation*} \left|\frac{1}{n^2}-0\right|=\frac{1}{n^2} \lt \epsilon. \end{equation*}
Rearranging this gives (and using the fact that \(\epsilon\) is positive)
\begin{equation*} \frac{1}{\epsilon}\lt n^2 \end{equation*}
Taking the square-root of both sides then tells us that we need
\begin{equation*} n \gt \frac{1}{\sqrt{\epsilon}} \end{equation*}
Therefore, we could take \(N\) to be any number greater than \(1/\sqrt{\epsilon}\text{.}\) In particular, we could take \(N\) to be the smallest integer greater than \(1/\sqrt{\epsilon}\text{,}\) which we denote by
\begin{equation*} \left\lceil 1/\sqrt{\epsilon} \right\rceil. \end{equation*}

6.6.19.

Scratchwork.
For a given \(\epsilon \gt 0\text{,}\) we want to find \(\delta \gt 0\) so that the following holds:
\begin{equation*} 0 \lt |x-0| \lt \delta \implies |f(x)-0| \lt \epsilon. \end{equation*}
Since \(x\neq0\text{,}\) the latter inequality is equivalent to
\begin{equation*} \left|6x\sin\left(\frac{1}{x}\right)\right| \lt \epsilon. \end{equation*}
We can simplify the inequality we’re working with by using the fact that \(|\sin(\theta)|\) is always bounded by \(1\text{.}\) This means that
\begin{equation*} \left|6x\sin\left(\frac{1}{x}\right)\right|\leq |6x|. \end{equation*}
You may have seen something like this argument given when discussing the “Squeeze theorem” or “Sandwich theorem” in a calculus course.
Then, if we have \(|6x| \lt \epsilon\text{,}\) we’d also have
\begin{equation*} \left|6x\sin\left(\frac{1}{x}\right)\right| \lt \epsilon. \end{equation*}
But \(|6x| \lt \epsilon\) is satisfied when
\begin{equation*} |x| \lt \frac{\epsilon}{6}. \end{equation*}
Therefore, we need \(\delta\leq \epsilon/6\text{.}\)
Now we can write up the formal proof.

6.6.20.

Scratchwork.
When the sequence \((x_n)_{n\in\mathbb{N}}=\left((-1)^n+\dfrac{1}{n}\right)_{n\in\mathbb{N}}\) does not converge to \(0\) it means that \((x_n)\) satisfies the negation of the definition of convergence with \(L=0\text{.}\) That is \((x_n) \) does not converge to \(0\) is equivalent to the statement
\begin{equation*} \exists \epsilon\gt 0, \forall N \in\mathbb{N}, \exists n\gt N, \left|\left((-1)^n+\dfrac{1}{n}\right)-0\right|\geq \epsilon. \end{equation*}
Observe that because of the \((-1)^n\) term in the sequence, it behaves slightly differently when \(n\) is even or odd.
  • For all even \(k \in \mathbb{N}\text{,}\) we have \(x_k=(-1)^k+\dfrac{1}{k}=1+\dfrac{1}{k}\gt 1\text{.}\)
  • While for all odd \(k \in \mathbb{N}\) with \(k\geq 3\text{,}\) \(x_k=(-1)^k+\dfrac{1}{k}=-1+\dfrac{1}{k}\leq -\dfrac{2}{3}\text{.}\)
Thus, for any integer \(k \geq 2\text{,}\) we know that \(|x_k-0| = |x_k| \geq \frac{2}{3}\text{.}\)
So take \(\epsilon = \dfrac{1}{2}\text{,}\) and then for any \(N \in \mathbb{N}\text{,}\) so that \(n \gt N\text{,}\) set \(n = \max\{N,2\}+1\) and we can make the proof work.

6.6.21.

Scratchwork.
Recall from the definition of a limit that the proof should go like this: We’re given some \(\epsilon\gt 0\text{.}\) We need to find some \(N\in\mathbb{N}\) so that whenever \(n\gt N\text{,}\) we have
\begin{equation*} \left|1-\frac{2}{n^2}-\frac{3}{n^3}-1\right| \lt \epsilon. \end{equation*}
Let’s unravel this inequality to see how large \(n\) needs to be. We want to rewrite this inequality so that’s in the form \(n\) greater than something; this something will tell us how large we need to take \(N\) so that the inequality above is satisfied.
Now since,
\begin{equation*} \left|\left(1-\frac{2}{n^2}-\frac{3}{n^3}\right)-1\right| = \left|-\frac{2}{n^2}-\frac{3}{n^3} \right| = \frac{2}{n^2}+\frac{3}{n^3} \end{equation*}
the inequality can be simplified to
\begin{equation*} \frac{2}{n^2}+\frac{3}{n^3} \lt \epsilon. \end{equation*}
Because we have two different terms with \(n\) in them, it may be difficult to put the inequality in the form we’d like, that is, \(n\) greater than something. What we can do instead is consider which of the factors \(1/n^2\) or \(1/n^3\) is larger - and then only working with the larger one, by bounding the smaller one by it.
Since \(n \in \mathbb{N}\) we know that \(n \geq 1\text{.}\) Dividing this inequality by \(n^3\) (which is positive), then gives us
\begin{equation*} \frac{1}{n^2} \geq \frac{1}{n^3} \end{equation*}
This means that
\begin{equation*} \frac{2}{n^2}+\frac{3}{n^3} \leq \frac{2}{n^2}+\frac{3}{n^2} =\frac{5}{n^2}. \end{equation*}
Therefore, if we show that
\begin{equation*} \frac{5}{n^2}\lt \epsilon, \end{equation*}
we’ll also have that
\begin{equation*} \frac{2}{n^2}+\frac{3}{n^3} \lt \epsilon. \end{equation*}
Now, we can take the square root of both sides of the inequality \(5/n^2\lt\epsilon\text{,}\) and use the reverse implication in the fact given in the question to say that
\begin{equation*} \frac{\sqrt{5}}{\sqrt{\epsilon}}\lt n. \end{equation*}
We can actually reverse all of these steps (so we use the forward implication in the given fact), to get the following implication:
\begin{equation*} \frac{\sqrt{5}}{\sqrt{\epsilon}}\lt n \implies \left|1-\frac{2}{n^2}-\frac{3}{n^3}-1\right| \lt \epsilon. \end{equation*}
This tells us that we can take \(N\) to be any natural number that is larger than \(\sqrt{5}/\sqrt{\epsilon}\text{.}\)
So we can now write up the proof.
But we have actually missed an opportunity to simplify things further here. Recall that by dividing the inequality \(n \geq 1\) by \(n^3\) we showed that \(\frac{1}{n^2} \geq \frac{1}{n^3}\) and used that to simplify the inequalities that we needed for our proof. We can simplify things further still by noting that by dividing the inequality \(n \geq 1\) by \(n^2\) we get
\begin{equation*} \frac{1}{n} \geq \frac{1}{n^2} \end{equation*}
So we can use that to show
\begin{equation*} \frac{2}{n^2}+\frac{3}{n^3} \leq \frac{2}{n^2}+\frac{3}{n^2} =\frac{5}{n^2} \leq \frac{5}{n}. \end{equation*}
Then it suffices to find \(n\) big enough so that
\begin{equation*} \frac{5}{n} \leq \epsilon \end{equation*}
which is quite a bit easier than the mucking around with square-roots we had to do before.

6.6.22.

Scratchwork.
  • (c): Let \(M\gt 0\text{.}\) We need to find some \(N\in \mathbb{N}\) so that whenever \(n\geq N\text{,}\) we have \(\sqrt{n}\geq M\text{.}\) But we know that \(\sqrt{n}\geq M\) if \(n\geq M^2\text{,}\) by Exercise 3.5.16. Therefore we can take \(N=M^2\text{.}\)
  • (d): We need to show that there is some \(M\gt0\text{,}\) so that for any \(N\in\mathbb{N}\text{,}\) there’s some \(n\geq N\) such that \((-1)^n\sqrt{n}\lt M\text{.}\) But if \(n\) is odd, then \((-1)^n\sqrt{n}\) is negative, so \((-1)^n\sqrt{n}\lt M\text{.}\) Thus we could actually take \(M\) to be any positive number, but for the purposes of the proof, we just need to fix a particular value of \(M\text{.}\) For example, we can take \(M=1\text{.}\)
  • (e): Let \(M\gt 0\text{.}\) We need to find some \(N\in \mathbb{N}\) so that whenever \(n\geq N\text{,}\) we have \(n^2-100n\geq M\text{.}\) Factoring the expression on the lefthand side of the inequality, we have
    \begin{equation*} n(n-100)\geq M. \end{equation*}
    Notice that if \(n-100\geq 1\text{,}\) then
    \begin{equation*} n(n-100)\geq n\cdot1=n. \end{equation*}
    So if \(n\geq 101\) and \(n\geq M\) we would have
    \begin{equation*} n^2-100n =n(n-100)\geq n\geq M, \end{equation*}
    as desired. So we need \(M\) to satisfy both \(n\geq M\) and \(n\geq 101\text{;}\) therefore we can take \(N=\max\{M,101\}\text{.}\)

6.6.23.

Scratchwork.
  1. By definition of the limit, we know that for any \(\epsilon\gt0\) there is some \(N_\epsilon\in\mathbb{N}\) such that
    \begin{equation*} n\geq N_\epsilon\implies |a_n-L|\lt\epsilon. \end{equation*}
    We would like to use the inequality \(|a_n-L|\lt\epsilon\) to bound \(|a_n|\) from above. We can do this by writing \(|a_n|=|L+(a_n-L)|\) and applying the triangle inequality. Indeed, we have
    \begin{equation*} |a_n|=|L+(a_n-L)|\leq |L|+|a_n-L|\lt |L|+\epsilon. \end{equation*}
    This inequality is true for all \(n\geq N_\epsilon\text{,}\) where \(N_\epsilon\) depends on our choice of \(\epsilon\gt0\text{.}\) We are free to choose \(\epsilon\text{,}\) so let’s take \(\epsilon=1\text{.}\) So for \(n\geq N_1\text{,}\) we have
    \begin{equation*} |a_n-L|\lt 1\implies |a_n|\lt |L|+1. \end{equation*}
    Now, we only have to bound
    \begin{equation*} |a_1|, |a_2|, \dots, |a_{N_1-1}|. \end{equation*}
    But there are only finitely many values here, so
    \begin{equation*} M_0=\max\{|a_1|, |a_2|, \dots, |a_{N_1-1}|\} \end{equation*}
    exists, and is a real number. So for any value of \(n\text{,}\) we know that \(|a_n|\leq M_0\) if \(1\leq n\lt N_1\text{,}\) or \(|a_n|\leq |L|+1\) if \(n\geq N_1\text{.}\) Therefore we may take \(M=\max\{M_0, |L|+1\}\text{.}\)
  2. We need to find a sequence \(\{a_n\}_{n\in\mathbb{N}}\) that is bounded, but does not converge. We can try a sequence that oscillates between two fixed numbers, say the sequence \(a_n=(-1)^n\text{.}\) This sequence is bounded, as \(|a_n|\leq 1\) for all \(n\text{.}\) Intuitively, this sequence shouldn’t converge since it doesn’t get arbitrarily close to a single number. Indeed, in Example 6.4.6, we saw that this sequence doesn’t converge to \(1\text{.}\) But we need to show a bit more than that for the purposes of this question; we need to show that for any \(L\in\mathbb{R}\text{,}\) the sequence doesn’t converge to \(L\text{.}\)
    To this end, we’ll fix \(L\in \mathbb{R}\text{.}\) To show that \(a_n\) doesn’t converge to \(L\text{,}\) we need to find some \(\epsilon\) so that the following is true:
    \begin{equation*} \forall N\in\mathbb{N}\; \exists n\geq N\; \text{ such that } \; |a_n-L|\geq\epsilon. \end{equation*}
    In Example 6.4.6, we saw that for \(L=1\) we could take \(\epsilon=1\text{,}\) because then for any odd \(n\text{,}\) \(|a_n-1|=|-1-1|\gt 1\text{.}\)

6.6.24.

Scratchwork.
We will prove part (c) by contrapositive. The contrapositive of the statement given in the question is
\begin{equation*} \lim_{n\to\infty}a_n\neq+\infty\implies (a_n)_{n\in\mathbb{N}} \text{ is bounded, or } (a_n)_{n\in\mathbb{N}} \text{ is not increasing.} \end{equation*}
So this statement is of the form \(P \implies (R \lor \neg Q) \text{.}\) Now \(Q\) is either true or false
  • If \(Q\) is false, then \(R \lor \neg Q\) is false, and so \(P \implies F\) is true. Easy!
  • While if \(Q\) is true, then \(R \lor \neg Q\) is true only when \(R\) is true, and so we have to show that \(P \implies R\) is true.
Also notice that this statement is of the same form as Exercise 4.3.1 (c), and so is logically equivalent to
\begin{equation*} \lim_{n\to\infty}a_n\neq+\infty\implies \left( (a_n)_{n\in\mathbb{N}} \text{ is increasing} \implies a_n \text{ is bounded.} \right) \end{equation*}
and so is of the form \(P \implies (Q \implies R)\text{.}\) Now, either \(Q\) is true or false:
  • If \(Q\) is false, then \(F \implies R\) is true, and so \(P \implies T\) is true. Easy!
  • On the other hand, if \(Q\) is true, then \(T \implies R\) is only true when \(R\) is true, so we really have to show that \(P \implies R\) is true.
We know one of the following is true: either \((a_n)_{n\in\mathbb{N}} \) is not increasing, or it is increasing. We will use these two cases to prove the contrapositive statement. In the first case, when \((a_n)_{n\in\mathbb{N}} \) is not increasing, the conclusion (that \((a_n)_{n\in\mathbb{N}} \) is bounded or not increasing) is automatically satisfied. Therefore we may focus on the second case, when \((a_n)_{n\in\mathbb{N}} \) is increasing. We need to show that the sequence is bounded (then the conclusion, \((a_n)_{n\in\mathbb{N}} \) is bounded or not increasing, will be true).
By assumption, we know that \(\displaystyle\lim_{n\to\infty}a_n\neq+\infty\text{.}\) By definition, this means there is some \(C\gt0\) such that for any \(n\in\mathbb{N}\text{,}\) there’s some \(m\geq n\) with \(a_m\lt C\text{.}\) But since \((a_n)_{n\in\mathbb{N}} \) is increasing, we know that \(a_n\leq a_m\lt C\text{.}\) While it’s not necessarily true that \(a_n\leq a_m\) for all \(n\text{,}\) since \(m\) depends on \(n\text{,}\) we do know that \(a_n\lt C\) for all \(n\text{.}\) Moreover, we also know that \(a_n\geq a_1\) for all \(n\in\mathbb{N}\text{,}\) as \(a_n\) is increasing. Combining these inequalities to get \(a_1\leq a_n\lt C\text{,}\) we can show that \(|a_n|\) is bounded, as desired.

6.6.25.

Scratchwork.
This particular distance function is known as the “discrete distance” or the “discrete metric”. Proving (a) is straight-forward by case analysis with either \(x=z\) or \(x \neq z\text{.}\)
Now, (b) looks like a complicated question since the distance that is given in the definition is not the standard distance function (where the distance between two points is defined as the absolute value of their difference). It also forces the reader to parse a new definition carefully and rework their understanding of the topic at hand.
Now, let’s try to see how we can prove this statememt.
First, we see that this is a conditional statement. This means that we can start by assuming the hypothesis and trying to show the conclusion. That is, assume that \((x_n)_{n\in\mathbb{N}}\) converges to \(L\text{,}\) and try to show that the set \(\set{ n\in\mathbb{N}: x_n\neq L }\) is finite.
Since \((x_n)_{n\in\mathbb{N}}\) converges to \(L\text{,}\) we know that
\begin{equation*} \forall \epsilon\gt 0, \exists N_0\in\mathbb{N}\text{ such that } \forall n\geq N_0, D(x_n,L)\lt \epsilon. \end{equation*}
Hence, since this is true for all \(\forall \epsilon\gt 0\text{,}\) we know that this should also work for \(\epsilon=1/2\) (any \(\epsilon\) strictly between \(0\) and \(1\) would work here). So, for \(\epsilon=1/2\text{,}\) we should be able to find an \(N_\epsilon\in\mathbb N\text{,}\) such that \(\forall n\gt N_\epsilon\text{,}\) we have \(D(x_n,L)\lt \epsilon=1/2\text{.}\)
Now, if we recall that the distance function \(D\) only takes values \(0\) or \(1\text{,}\) this means that \(D(x_n,L)=0\lt\epsilon=1/2\lt 1\text{.}\)
We see that this is what we needed to show, since this means that \(\forall n\gt N_\epsilon\text{,}\) we have \(x_n=L\text{.}\) So the only \(x_n \neq L\) must occur for \(n \leq N_\epsilon\text{,}\) and that is a finite set.
Let’s clean this up and write in a proof.

7 Induction
7.3 Exercises

7.3.2.

Scratchwork.
When \(n=k\) the statement is
\begin{equation*} k! \leq k^k \end{equation*}
and we will need to prove that when \(n=k+1\)
\begin{equation*} (k+1)! \leq (k+1)^(k+1) \end{equation*}
Notice that \((k+1)! = k! \cdot (k+1)\text{,}\) and so we can make the first statement look like the second by multiplying by \((k+1)\text{.}\)

7.3.3.

Scratchwork.
Since we want to show that the inequality \(n!\gt 3^n\) is true for all \(n\geq 7\text{,}\) we can try to understand how it behaves for any natural number. We can do that by starting to look at the statement for different values of \(k\in\mathbb N\text{,}\) and try to develop a pattern.
We see that for \(n=1\text{,}\) the inequality becomes \(1\gt 3\text{,}\) which is not true. Moreover, if we check couple more numbers, we get,
\begin{align*} n=2 \amp \amp 2\gt 9 \amp \amp \text{False},\\ n=3 \amp \amp 6\gt 27 \amp \amp \text{False},\\ n=4 \amp \amp 24\gt 81 \amp \amp \text{False},\\ n=5 \amp \amp 120\gt 243 \amp \amp \text{False},\\ n=6 \amp \amp 720\gt 729 \amp \amp \text{False},\\ n=7 \amp \amp 5040\gt 2187 \amp \amp \text{True},\\ n=8 \amp \amp 40320\gt 6561 \amp \amp \text{True},\\ n=9 \amp \amp 362880\gt 19683 \amp \amp \text{True}. \end{align*}
After \(n=7\text{,}\) we see that the left hand side grows much faster than the right hand side of the inequlity. This suggests that the statement shoul be true. Now, we can try to show it using induction.

7.3.5.

Scratchwork.
We want to prove this statement by induction. As written, we want to prove something about all \(m\in\mathbb{N}\text{,}\) with \(m\) even. In order to prove this by induction, however, we need reformulate the statement into something that we want to prove for all natural numbers instead. We can do this by writing \(m=2n\) for \(n\in\mathbb{N}\text{.}\) Then we need to prove that \(8\) divides \(3^{2n}-1\text{.}\) We can prove this by induction on \(n\text{.}\)
  • Base case: Take \(n=1\text{.}\) Then \(3^{2\cdot n}-1 =9-1=8, \text{,}\) so the result holds.
  • Inductive step: Assume that
    \begin{equation*} 8\mid 3^{2k}-1. \end{equation*}
    We want to show that
    \begin{equation*} 8\mid 3^{2(k+1)}-1. \end{equation*}
    Our assumption implies that there is some \(\ell\) so that
    \begin{equation*} 3^{2k}-1 = 8 \ell \end{equation*}
    Rephrasing this we have \(3^{2k} = 8\ell+1\text{.}\) Notice that this implies that
    \begin{equation*} 3^{2k+2} = 72\ell + 9 \end{equation*}
    and thus
    \begin{equation*} 3^{2k+2}-1 = 72\ell + 8 = 8(9\ell + 1) \end{equation*}
    giving us the divisibility result we need.
  • Inductive step — another way: Another way to approach the inductive step is to consider the difference between
    \begin{equation*} (3^{2k}-1) \qquad \text{and} \qquad (3^{2k+2}-1). \end{equation*}
    If that difference is a multiple of 8 then we can infer the result we need. So again, we assume that \(8 \mid 3^{2k}-1\text{,}\) and then consider the difference:
    \begin{equation*} (3^{2k+2}-1) - (3^{2k}-1) = 3^{2k+2} - 3^{2k} = 9 \cdot 3^{2k} - 3^{2k} = 8\cdot 3^{2k}. \end{equation*}
    Then since \(8 \mid 3^{2k}-1\) we can write it as \(8\ell\) for some integer \(\ell\text{,}\) and so
    \begin{equation*} (3^{2k+2}-1) - \underbrace{(3^{2k}-1)}_{=8\ell} = 8 \cdot 3^{2k} \end{equation*}
    and so, with a little rearranging:
    \begin{equation*} (3^{2k+2}-1) = 8 \ell + 8 \cdot 3^{2k}. \end{equation*}
    So we see that \(3^{2(k+1)}-1\) is divisible by \(8\text{.}\)
Now we can write the formal proof up:

7.3.6.

Scratchwork.
We want to prove this statement by induction on \(n\text{.}\)
  • Base case: Take \(n=2\text{.}\) Then the distributive law gives the result in this case.
  • Inductive step: Assume that the result holds for \(n=k\text{,}\) and let \(a,b_1,\dots,b_k,b_{k+1}\) be real numbers. The inductive step tells us that
    \begin{equation*} a\cdot(b_1+b_2+\dots+b_k)=a\cdot b_1+a\cdot b_2+\dots+a\cdot b_k, \end{equation*}
    which we want to use to show that
    \begin{equation*} a\cdot(b_1+b_2+\dots+b_k+b_{k+1})=a\cdot b_1+a\cdot b_2+\dots+a\cdot b_k+a\cdot b_{k+1}. \end{equation*}
    We want to consider the sum \(b_1+b_2+\dots+b_k+b_{k+1}\) as the sum of two elements, \(b_1+b_2+\dots+b_k\) and \(b_{k+1}\text{.}\) Then the distributive law tells us that
    \begin{equation*} a\cdot((b_1+b_2+\dots+b_k)+b_{k+1})=a\cdot(b_1+b_2+\dots+b_k)+a\cdot b_{k+1}. \end{equation*}
    Then we can apply the inductive hypothesis to obtain the desired result.
Now we can write the formal proof up:

7.3.7.

Scratchwork.
We need to prove the statement by induction. Let’s take a look at the base case and the inductive step.
  • Base case: \(n=1\text{.}\) Then \(2^n=2= 2n\text{,}\) so \(2^n\geq 2n\text{.}\)
  • Inductive step: Suppose that the inequality holds for \(n=k\text{,}\) so that \(2^k\geq 2k\text{.}\) We need to show that \(2^{k+1}\geq 2(k+1)\text{.}\)
    Note that if we multiply \(2^k\geq 2k\) by \(2\text{,}\) we’ll end up with \(2^{k+1}\) on the right-hand side, which is the right-hand side of the inequality we want to show. So, let’s do that to obtain \(2^{k+1}\geq 4k\text{.}\)
    Now, we just need to show that \(4k\geq 2(k+1)\text{,}\) and then we’ll have \(2^{k+1}\geq 2(k+1)\text{,}\) as desired. Let’s try to simplify the inequality \(4k\geq 2(k+1)\) to see why it may be true. Rewrite the inequality as
    \begin{equation*} 4k\geq 2k+2. \end{equation*}
    Subtracting \(2k\) from both sides of the inequality, we have \(2k\geq 2\text{,}\) which after dividing by \(2\text{,}\) gives \(k\geq1\text{.}\) But we know that \(k\geq1\text{,}\) since \(k\in\mathbb{N}\text{.}\)
    We can reverse all of these steps to show that \(k\geq 1\) implies \(4k\geq 2(k+1)\text{.}\) When we write up the formal proof, it’s important to start with what we know, namely, that \(k\geq 1\text{,}\) in order to show the desired inequality, that \(4k\geq 2(k+1)\text{.}\)
Now we’re ready to write up the proof.

7.3.8.

Scratchwork.
We want to prove this statement by induction on \(n\text{.}\)
  • Base case: Take \(n=0\text{.}\) Then \(2^{2n+1}+3^{2n+1} =2+3=5, \text{,}\) so the result holds.
  • Inductive step: Assume that
    \begin{equation*} 5\mid \left(2^{2k+1}+3^{2k+1}\right). \end{equation*}
    We want to show that
    \begin{equation*} 5\mid \left( 2^{2(k+1)+1}+3^{2(k+1)+1} \right). \end{equation*}
    The assumption tells us that there is some \(\ell\in\mathbb{Z}\) such that
    \begin{equation*} 2^{2k+1}+3^{2k+1} = 5\ell. \end{equation*}
    Note that
    \begin{equation*} 2^{2(k+1)+1}+3^{2(k+1)+1} =4\cdot2^{2k+1}+9\cdot 3^{2k+1}=4(2^{2k+1}+3^{2k+1})+5\cdot 3^{2k+1} \end{equation*}
    and so by the inductive hypothesis
    \begin{equation*} 2^{2(k+1)+1}+3^{2(k+1)+1} =4\cdot 5\ell+5\cdot 3^{2k+1} =5(4\ell+3^{2k+1}). \end{equation*}
    Therefore \(2^{2(k+1)+1}+3^{2(k+1)+1}\) is divisible by \(5\text{.}\)
  • Another way to prove the inductive step is to consider the difference
    \begin{equation*} 2^{2(k+1)+1}+3^{2(k+1)+1}-(2^{2k+1}+3^{2k+1}) \end{equation*}
    and show that it is divisible by \(5\text{.}\) To this end, let’s simplify that difference:
    \begin{align*} 2^{2(k+1)+1}+3^{2(k+1)+1}-(2^{2k+1}+3^{2k+1}) \amp =3^{2(k+1)+1}-3^{2k+1}+2^{2(k+1)+1}-2^{2k+1}\\ \amp =3^2\cdot 3^{2k+1}-3^{2k+1}+2^2\cdot 2^{2k+1}-2^{2k+1}\\ \amp =8\cdot 3^{2k+1}+3\cdot2^{2k+1} \end{align*}
    We want to rewrite this expression in terms of \(2^{2k+1}+3^{2k+1}\text{,}\) so that we can use the inductive hypothesis. We rewrite the expression above as
    \begin{equation*} 8\cdot 3^{2k+1}+3\cdot2^{2k+1} = 8\cdot (3^{2k+1}+2^{2k+1})-8\cdot 2^{2k+1}+3\cdot2^{2k+1} \end{equation*}
    This is equal to
    \begin{equation*} 8\cdot (3^{2k+1}+2^{2k+1})-5\cdot2^{2k+1} \end{equation*}
    which is divisible by \(5\text{,}\) since the second term is divisible by \(5\text{,}\) and the first is by the inductive hypothesis.
Now we can write the formal proof.

7.3.9.

Scratchwork.
The base case of our induction will be \(n=2\text{.}\) For this part, we’ll need to show that the sum of two rational numbers is rational. For the inductive step, we’ll suppose that any sum of \(k\) rational numbers is rational, and we need to show that if \(x_1,x_2,\dots,x_{k+1}\in\mathbb{Q}\text{,}\) then their sum lies in \(\mathbb{Q}\text{.}\) The inductive hypothesis tells us that \(x_1+x_2+\dots+x_k\in\mathbb{Q}\text{.}\) But then \(x_1+x_2+\dots+x_k+x_{k+1}\) is really the sum of two rational numbers, and hence rational itself, as we already proved this in the base case.

7.3.10.

Scratchwork.
Here we see that we have a statement we want to show is true for all \(n\in\mathbb N\text{.}\) In the previous chapters we have seen that to show such statements we assume that \(n\in\mathbb N\) and try to show the conclusion. Moreover, since the hypothesis is broad, we would probably need to look at cases. But, in this question, it is not clear what the cases should be. Instead, we can also apply mathematical induction. We see that this equation doesn’t seem to give us a formula, but instead gives us two expressions. However, we know what the right hand side is equal to.
We know that \(\displaystyle{\sum\limits_{k=1}^m k}=\dfrac{m(m+1)}{2}\) (see Result 7.1.8). This means that what we want to show now becomes
\begin{equation*} \sum\limits_{k=1}^{n} k^3=\frac{n^2(n+1)^2}{4}. \end{equation*}
Now, we see that the how we can use induction becomes more clear.

7.3.11.

Scratchwork.
In this question we want to prove that an inequality is true for all \(n\in\mathbb N\text{.}\) This suggests that induction may be a good method. However, in general, proving inequalities using induction may be a little harder than proving equalities or formulas. The reason for that is the fact that to prove the inductive step, we may need to manipulate the inequality, add or subtract terms, etc. We should also be very careful to make sure that our logic flows correctly in the proof of the inductive step; start from the assumption and reach the conclusion.
If we look at this question, and try to understand how we can handle the inductive step, we would assume that
\begin{equation*} \sum\limits_{j=1}^n j^3\gt\frac{1}{4}n^4, \end{equation*}
and try to show
\begin{equation*} \sum\limits_{j=1}^{n+1} j^3\gt\frac{1}{4}(n+1)^4. \end{equation*}
We see that (based on our assumption)
\begin{equation*} \sum\limits_{j=1}^{n+1} j^3=\sum\limits_{j=1}^n j^3+(n+1)^3\gt\frac{1}{4}n^4+(n+1)^3. \end{equation*}
But, \(\frac{1}{4}n^4+(n+1)^3\neq \frac{1}{4}(n+1)^4\text{.}\)
To complete the proof, we need to show that
\begin{equation*} \frac{1}{4}n^4+(n+1)^3\gt \frac{1}{4}(n+1)^4=\frac{1}{4}n^4+n^3+\frac{3}{2}n^2+n+\frac{1}{4}, \end{equation*}
which requires just a little careful algebra.

7.3.12.

Scratchwork.
Let \(r \neq 1\text{,}\) and then we need to prove the result
\begin{equation*} \sum_{i=0}^nr^i = \frac{1-r^{n+1}}{1-r} \end{equation*}
for all \(n\in\set{0,1,2,\dots}\text{.}\) Notice that since \(r \neq 1\) the expression is okay; we do not divide by zero. For the base case, we need to show that the result holds for \(n=0\text{,}\) while for the inductive step, we need to assume that the result is true for \(n=k\text{,}\) and then show that the result also holds for \(n=k+1\text{.}\)
  • The base case is true, since
    \begin{equation*} \sum_{i=0}^0r^i=r^0=1=\frac{1-r}{1-r}=\frac{1-r^{0+1}}{1-r}. \end{equation*}
  • For the inductive step, we assume that
    \begin{equation*} \sum_{i=0}^kr^i = \frac{1-r^{k+1}}{1-r} \end{equation*}
    and we need to show that
    \begin{equation*} \sum_{i=0}^{k+1}r^i = \frac{1-r^{(k+1)+1}}{1-r}. \end{equation*}
    Notice that we can obtain the left-hand side of the equation above by adding \(r^{k+1}\) to the left-hand side of the inductive hypothesis. Then
    \begin{equation*} \sum_{i=0}^{k+1}r^i=\sum_{i=0}^kr^i+r^{k+1}=\frac{1-r^{k+1}}{1-r}+r^{k+1}, \end{equation*}
    by the inductive hypothesis. Now we just need to simplify the right-hand side of this equation:
    \begin{equation*} \frac{1-r^{k+1}}{1-r}+r^{k+1}=\frac{1-r^{k+1}+r^{k+1}(1-r)}{1-r}=\frac{1-r^{k+2}}{1-r}, \end{equation*}
    and we end up with
    \begin{equation*} \sum_{i=0}^{k+1}r^i = \frac{1-r^{(k+1)+1}}{1-r} \end{equation*}
    which is what we wanted to show.

7.3.13.

Scratchwork.
  1. \(f(x) = x^n\) for \(n\in\mathbb{N}\) with \(0\leq k\leq n\text{.}\)
    We compute the first few derivatives:
    \begin{align*} f^{(1)}(x) \amp = nx^{n-1}\\ f^{(2)}(x) \amp = n(n-1)x^{n-2}\\ f^{(3)}(x) \amp = n(n-1)(n-2)x^{n-3}\\ f^{(4)}(x) \amp = n(n-1)(n-2)(n-3)x^{n-4} \end{align*}
    A reasonable guess for the \(k^{\text{th}}\) derivative would be
    \begin{equation*} f^{(k)} (x) = n(n-1)\dots (n-(k-1))x^{n-k} = \frac{n!}{(n-k)!}x^{n-k} \end{equation*}
  2. \(g(x)=x^{-n}\) for \(n\in \mathbb{N}\text{.}\)
    \begin{align*} g^{(1)}(x) \amp = -nx^{-n-1}\\ g^{(2)}(x) \amp = -n(-n-1)x^{-n-2}\\ g^{(3)}(x) \amp = -n(-n-1)(-n-2)x^{-n-3}\\ g^{(4)}(x) \amp = -n(-n-1)(-n-2)(-n-3)x^{-n-4} \end{align*}
    A reasonable guess for the \(k^{\text{th}}\) derivative would be
    \begin{equation*} g^{(k)}(x) = (-1)^k n(n+1)\dots (n+(k-1))x^{n-k} =\frac{(-1)^k (n+(k-1))!}{(n-1)!}x^{-n-k} \end{equation*}
  3. \(h(x)=\frac{1}{\sqrt{9-2x}} = (9-2x)^{-1/2}\text{.}\) This one is a little harder:
    \begin{align*} h^{(1)}(x) \amp = -\frac{1}{2}(-2)(9-2x)^{-3/2} = (9-2x)^{-(2(1)+1)/2}\\ h^{(2)}(x) \amp = \frac{-3}{2}(-2)(9-2x)^{-5/2} = 3(9-2x)^{-(2(2)+1)/2}\\ h^{(3)}(x) \amp = (3)\frac{-5}{2}(-2)(9-2x)^{-7/2} = 3\cdot 5(9-2x)^{-(2(3)+1)/2}\\ h^{(4)}(x) \amp = (3\cdot 5)\frac{-7}{2} (-2)(9-2x)^{-9/2} = 3\cdot 5 \cdot 7 (9-2x)^{-(2(4)+1)/2} \end{align*}
    A reasonable guess for the \(k^{\text{th}}\) derivative would be
    \begin{equation*} h^{(k)}(x) = (3)(5) \dots (2(k-1)+1)) (9-2x)^{-(2k+1)/2} = (2k-1)!! (9-2x)^{-(2k+1)/2} \end{equation*}
    where we have used the double factorial \(\ell !!\)
    \begin{equation*} \ell !! = \begin{cases} \ell \cdot (\ell -2) \cdots 2 \amp \text{when } \ell \text{ is even} \\ \ell \cdot (\ell -2) \cdots 3 \cdot 1 \amp \text{when } \ell \text{ is odd} \end{cases}. \end{equation*}

7.3.14.

Scratchwork.
As scratchwork for this problem, we work through the computation we will have to do after we adopt the inductive hypothesis. That is, we want to work through the \(n=j+1\) case after assuming that the \(n=j\) case is true.
\begin{align*} \sum_{k=1}^{j+1} k^2 \amp = (j+1)^2 + \sum_{k=1}^{j} k^2 \end{align*}
By the inductive hypothesis,
\begin{align*} \amp = (j+1)^2 + \frac{j(j+1)(2j+1)}{6}. \end{align*}
Expanding the expression and giving both terms a common denominator yields
\begin{align*} \amp =\frac{(6j^2+12j+6) + (2j^3+3j^2+j)}{6}. \end{align*}
Grouping like terms,
\begin{align*} \amp = \frac{2j^3+9j^2+13j+6}{6} \end{align*}
We now factor the top of the expression to get
\begin{align*} \amp = \frac{(j+1)(j+2)(2j+3)}{6} \end{align*}
Finally, we rearrange into the desired form
\begin{align*} \amp = \frac{(j+1)((j+1)+1)(2(j+1)+1)}{6}. \end{align*}
The above method works, but, in line (2), we could also notice that \(j+1\) is a common factor, rather than expanding the cubic. That computation looks like:
\begin{align*} \sum_{k=1}^{j+1} k^2 \amp = (j+1)^2 + \sum_{k=1}^{j} k^2 \end{align*}
By the inductive hypothesis,
\begin{align*} \amp = (j+1)^2 + \frac{j(j+1)(2j+1)}{6}. \end{align*}
Factoring out \(j+1\) and giving both terms a common denominator yields
\begin{align*} \amp =\frac{(j+1)(6(j+1)+j(2j+1))}{6}. \end{align*}
Simplifying,
\begin{align*} \amp = \frac{(j+1)(2j^2+7j+6)}{6} \end{align*}
We now factor the top of the expression to get
\begin{align*} \amp = \frac{(j+1)(j+2)(2j+3)}{6} \end{align*}
Finally, we rearrange into the desired form
\begin{align*} \amp = \frac{(j+1)((j+1)+1)(2(j+1)+1)}{6}. \end{align*}

7.3.17.

Scratchwork.
  1. We’ll prove this by induction on \(n\text{.}\)
    • Base case: take \(n=0\text{.}\) Then
      \begin{equation*} \lim_{x\to\infty}x^ne^{-x}=\lim_{x\to\infty}e^{-x}=0. \end{equation*}
    • Inductive step: Assuming
      \begin{equation*} \lim_{x\to\infty}x^ke^{-x}=0, \end{equation*}
      we want to show that
      \begin{equation*} \lim_{x\to\infty}x^{k+1}e^{-x}=0. \end{equation*}
      Rewrite the limit as
      \begin{equation*} \lim_{x\to\infty}\frac{x^{k+1}}{e^{x}} \end{equation*}
      Then both the numerator and the denominator go to infinity as \(x\to\infty\text{.}\) Then we can use l’Hôpital’s rule to obtain
      \begin{equation*} \lim_{x\to\infty}\frac{x^{k+1}}{e^{x}}=\lim_{x\to\infty}\frac{(k+1)x^{k}}{e^{x}}. \end{equation*}
      Now we have a limit that almost looks like the one from the inductive hypothesis; we just need to pull out the constant factor of \(k+1\text{,}\) and then apply the inductive hypothesis:
      \begin{equation*} \lim_{x\to\infty}\frac{x^{k+1}}{e^{x}}=(k+1)\lim_{x\to\infty}\frac{x^{k}}{e^{x}}=(k+1)\cdot0=0. \end{equation*}
  2. We’ll prove this by induction on \(n\text{.}\)
    • Base case: take \(n=0\text{.}\) Then
      \begin{equation*} \int_0^\infty x^ne^{-x}\mathrm{d}x =\int_0^\infty e^{-x}\mathrm{d}x \end{equation*}
      Evaluating this integral yields
      \begin{equation*} \int_0^\infty e^{-x}\mathrm{d}x = \lim_{t\to\infty}(-e^{-x})\Big]_0^t = 1. \end{equation*}
      And since \(1=0!\) the base case holds.
    • Inductive step: Suppose that
      \begin{equation*} k! = \int_0^\infty x^ke^{-x}\mathrm{d}x. \end{equation*}
      We want to use this equation in order to show that
      \begin{equation*} (k+1)! = \int_0^\infty x^{k+1}e^{-x}\mathrm{d}x. \end{equation*}
      We can connect the integral in the equation for \((k+1)!\) to the integral in the equation for \(k!\) by using integration by parts, because \(\frac{d}{\mathrm{d}x}x^{k+1}=(k+1)x^k\text{.}\) Indeed, by integration by parts,
      \begin{align*} \int_0^\infty x^{k+1}e^{-x}\mathrm{d}x \amp =(-x^{k+1}e^{-x})\Big]_0^\infty +(k+1)\int_0^\infty x^ke^{-x}\mathrm{d}x\\ \amp =\lim_{t\to\infty} \left(-t^{k+1}e^{-t}\right)+0^{k+1}e^{-0} +(k+1)\int_0^\infty x^ke^{-x}\mathrm{d}x \end{align*}
      We then use our result from (a) to evaluate that limit, giving
      \begin{align*} \amp = 0 +(k+1)\int_0^\infty x^ke^{-x}\mathrm{d}x \end{align*}
      Now we’re left with \((k+1)\) times the integral that we’ve assumed is equal to \(k!\) in the inductive hypothesis. But this gives exactly the equation that we wanted to show!

7.3.20.

Scratchwork.
Let’s try making this work with \(F_7 = 13\text{.}\) That is, we want to show that \(F_7 \mid F_7n\text{.}\) Se we’ll assume that \(F_7 \mid F_{7k}\) and then we’ll try to show that this implies that \(F_7 \mid F_{7k+7}\text{.}\) Now, from the definition of the Fibonacci numbers, we can write \(F_{7k+7}\) in terms of \(F_{7k+6}\) and \(F_{7k+5}\text{:}\)
\begin{equation*} F_{7k+7} = F_{7k+6}+F_{7k+5} \end{equation*}
and then similarly expand \(F_{7k+6} = F_{7k+5}+F_{7k+4}\text{,}\) so:
\begin{equation*} F_{7k+7} = 2F_{7k+5}+F_{7k+4} \end{equation*}
and keep going:
\begin{align*} F_{7k+7} \amp = 2F_{7k+5}+F_{7k+4} \amp \text{expand } F_{7k+5}\\ \amp= 3F_{7k+4} + 2 F_{7k+3} \amp \text{expand } F_{7k+4}\\ \amp= 5F_{7k+3} + 3 F_{7k+2} \amp \text{expand } F_{7k+3}\\ \amp= 8F_{7k+2} + 5 F_{7k+1} \amp \text{expand } F_{7k+2}\\ \amp= 13F_{7k+1} + 8 F_{7k} \end{align*}
Now since, by assumption \(13=F_{7} \mid F_{7k}\text{,}\) we are done.
Notice that each time we expand and regroup the terms we get another Fibonacci number. So, let us try to write this more explicitly in terms of Fibonacci numbers:
\begin{align*} F_n \amp = 1 \cdot F_{n-1} + 1 \cdot F_{n-2}\\ \amp = F_2 \cdot F_{n-1} + F_1 \cdot F_{n-2}\\ \amp = F_3 \cdot F_{n-2} + F_2 \cdot F_{n-3} \\ \amp = F_4 \cdot F_{n-3} + F_3 \cdot F_{n-4} \\ \amp = F_5 \cdot F_{n-4} + F_4 \cdot F_{n-5} \end{align*}
So we might try to show, more generally, that
\begin{equation*} F_n = F_{\ell} \cdot F_{n-\ell+1} + F_{\ell-1} \cdot F_{n-\ell} \end{equation*}
or equivalently (after substituting \(n \mapsto n+\ell\))
\begin{equation*} F_{n+\ell} = F_{\ell} \cdot F_{n+1} + F_{\ell-1} \cdot F_{n}. \end{equation*}
Setting \(n=7k, \ell=7\) in this gives us
\begin{equation*} F_{7k+7} = F_{7} \cdot F_{7k+1} + F_{6} \cdot F_{7k} \end{equation*}
which is precisely what we need.

7.3.23.

Scratchwork.
As scratchwork, we want to try to break the \(n+1\) case into something involving the previous cases.
Rather than trying to factor \(\alpha^{n+1}+\frac{1}{\alpha^{n+1}}\text{,}\) we attempt to get to it by multiplying the previous cases. We try to multiply the \(n=1\) and \(n\) cases because this will give a power of \(n+1\)
\begin{equation*} \left(\alpha^n+\frac{1}{\alpha^n}\right)\left(\alpha + \frac{1}{\alpha}\right) = \alpha^{n+1}+\frac{1}{\alpha^{n+1}} +\alpha^{n-1}+\frac{1}{\alpha^{n-1}} \end{equation*}
We now have the term we want in right hand side! Rearranging, we see
\begin{equation*} \alpha^{n+1}+\frac{1}{\alpha^{n+1}} = \left(\alpha^n+\frac{1}{\alpha^n}\right)\left(\alpha + \frac{1}{\alpha}\right) -\left(\alpha^{n-1}+\frac{1}{\alpha^{n-1}}\right). \end{equation*}
Since we can write the \((n+1)^{\text{st}}\) case in terms of the \(n^{\text{th}}\) and \((n-1)^{\text{st}}\) cases, we should make sure to have two consecutive cases (e.g. \(n=0,1\)) in the base case.
Notice that we are able to use an induction argument with two parts to the base case, since this allows us to have two parts to the inductive hypothesis. Another way of proving this is to notice that for the \(n+1\) case we need to reference more than just the previous case, so we can use proof by strong induction.

7.3.24.

Scratchwork.
This question tell us that from a natural number \(n\text{,}\) we can divide out the biggest possible power of \(3\text{,}\) say \(3^m\text{,}\) so that \(3\nmid \left( \frac{n}{3^m} \right)\text{.}\) Since this is a statement that should hold for all natural numbers, induction may be useful method to use. But, because of the nature of the problem, i.e. since we are trying to factor out as many powers of \(3\) as possible, we see that it may be hard to get from \(n=k\) to \(n=k+1\text{,}\) but it may be easier to get from \(n=\frac{(k+1)}{3}\) to \(n=k+1\) assuming that \(\frac{(k+1)}{3}\in\mathbb Z\text{.}\) This suggests that we may need to use strong induction rather than regular induction.

7.3.26.

Scratchwork.
We want to prove this statement by induction on \(n\text{.}\) Since \(a_0\) and \(a_1\) are not given by the recurrence relation, we’ll prove that the formula is satisfied for \(n=0\) and \(n=1\) in the base case. When \(n\geq 2\text{,}\) we have to use the recurrence relation \(a_n = a_{n-1}+6a_{n-2}\) to prove that the formula holds. We can do this by plugging in the formula for \(a_{n-1}\) and \(a_{n-2}\text{.}\) Consequently, for the inductive hypothesis, we need to rely on the statement for \(n-1\) and \(n-2\text{,}\) not just \(n-1\text{.}\) Therefore we need to use strong induction.

7.3.27.

Scratchwork.
You may have already seen how to write a natural number in binary:
  • We take the natural number,
  • find the biggest power of \(2\) that is smaller than that number,
  • subtract that power of \(2\) from the number,
  • and keep doing this till you end up with \(0\) or \(1\text{.}\)
Here we see that, even though this method is practically useful, this doesn’t give us the proof. Since there are some gaps in the argument. For example, how can we pick that “biggest factor of \(2\)”, or what does it mean to “keep doing till we end up with \(0\) or \(1\)”?
This is a good use of induction methods since induction takes care of that “keep going till…” portion of the process.
Now, we may want to use regular induction in this question. If we think about how we can prove the inductive step, to get from \(n\) to \(n+1\text{,}\) we would assume \(n=c_m2^m+c_{m-1}2^{m-1}+\cdots+c_12+c_0\) for some \(m\geq 0\) and constants \(c_0,c_1,c_2,\dots, c_m\in\set{0,1} \text{.}\) Then, we see \(n+1=c_m2^m+c_{m-1}2^{m-1}+\cdots+c_12+c_0+1\text{.}\) But then, we have to have \(2m\) different cases depending on whether or not \(c_i\) is \(0\) or not, for \(0\leq i\leq m\text{.}\) We see that this has a very similar “keep going till...” type of problem. Instead, we can use strong induction and the fact that every natural number is even or odd.
Notice that in our argument we are implicitly making use of the distributive law
\begin{equation*} 2 \cdot (a_0+a_1+a_2+\cdots+a_n) = 2\cdot a_0 + 2\cdot a_1 +2\cdot a_2 +\cdots +2\cdot a_n \end{equation*}
Now normally, we state the distributive law as
\begin{equation*} a \cdot(b+c) = a\cdot b + a\cdot c, \end{equation*}
that is, we show how we distribute the produce over two summands, rather than \(n\) summands. So we should really prove that it extends in this way. That is precisely Exercise 7.3.6.

8 Return to sets
8.6 Exercises

8.6.2.

Scratchwork.
The sets \(F\) and \(G\) may look like regular sets in \(\mathbb R^2\text{,}\) but we see that in both cases, the second item of each ordered pair is given as a function of the first coordinate. This suggests that \(F\) and \(G\text{,}\) indeed, can be seen as representing the graphs of the functions \(f(x)=x^2-3x+2\) and \(g(x)=x+2\) respectively. So, finding the intersection of these sets is the same as finding the intersection of these two graphs. Let’s see how that works in the proof.
Any element in \(F\) is of the form \((x,x^2-3x+2)\) and any element of \(G\) is of the form \((z,z+2)\text{.}\) So if an element is in the intersection we must have that
\begin{equation*} (x,x^2-3x+2) = (z,z+2) \end{equation*}
and so (since these are ordered pairs) \(x=z\) and \(x^2-3x+2 = z+2\text{.}\) Combining these we get the equation
\begin{equation*} x^2-3x+2=x+2 \end{equation*}
and so
\begin{equation*} x^2-4x=0 \end{equation*}
which implies that \(x=0,4\text{.}\) So the points in the intersection are \(\set{(0,2), (4,6)}\text{.}\)
But, we still have to prove that
\begin{equation*} F \cap G = \set{(0,2), (4,6)} \end{equation*}
We do this in two ways, first we prove a sequence of set-equalities, and then second we show that each side is a subset of the other.

8.6.3.

Scratchwork.
At a first glance, this statement looks like a true statement. It is because it sounds like it is saying: “if I remove elements of \(C\) from \(B\) and then add them back again, I get back to \(B\)”. But unfortunately it is not exactly what it says. What it says is: “if I remove all elements of \(C\) from \(B\) and add all the elements of \(C\) to it, then you get back to \(B\)”. Now, we see that this doesn’t have to be true since the set \(C\) may contain element that were not in \(B\) in the first place. So those elements wouldn’t have been removed when we took the difference, and would be added on top when we took the union. So, as long as the set \(C\) has an element that is not in \(B\text{,}\) we will have a counterexample. Let’s see how we can create such a counterexample for this statement.

8.6.4.

Scratchwork.
We want to prove that \(\set{x^a \st a\in\mathbb{Q} }=\set{ y^a\st a\in\mathbb{Q} }\) given that \(x, y \in\mathbb{R}\) and \(k \in\mathbb N\) satisfying, \(x,y \gt 0\) and \(x^k = y\text{.}\) At first glance it may look like a difficult question since it has many variables. But, we can simplify this statement and have a better intuition by choosing certain values of \(x\) and \(y\text{.}\)
For example, we can take \(x=\sqrt{3}\) and \(y=3\text{,}\) i.e. \(k=2\text{.}\) Then we see that the statement becomes:
\begin{equation*} \set{\left(\sqrt{3}\right)^a\st a\in\mathbb{Q} } = \set{\left(3^{\frac{1}{2}}\right)^a\st a\in\mathbb{Q} } = \set{3^{\frac{a}{2}}\st a\in\mathbb{Q} } = \set{ 3^b\st b\in\mathbb{Q} }. \end{equation*}
We see that the set on the left and the one on the right are the same sets since for every \(a\in\mathbb Q\text{,}\) we have \(b\in\mathbb Q\) such that \(b=\frac{a}{2}\) and for every \(b\in\mathbb Q\text{,}\) we have \(a\in\mathbb Q\) such that \(a=2b\text{.}\) This definitely does not work if we were forced to take integer \(a,b\text{.}\)
Now if we replace \(\sqrt 3\) and \(3\) with \(x\) and \(y\) with \(x,y \gt 0\) and \(x^k = y\text{,}\) same argument would still hold, but instead of \(k=2\text{,}\) we would have some other \(k\text{.}\) Namely,
\begin{equation*} \set{x^a\st a\in\mathbb{Q} } = \set{\left(y^{\frac{1}{k}}\right)^a\st a\in\mathbb{Q} } = \set{y^{\frac{a}{k}}\st a\in\mathbb{Q} } = \set{ y^b\st b\in\mathbb{Q} }. \end{equation*}
We can now write this neatly up as a proof.

8.6.5.

Scratchwork.
We see that this is a “prove or disprove” question. This means that we need to figure out whether it is true or not.
The statement claims that every element that is both in the sets \(\set{x \in\mathbb{Z} : m\mid x}\) and \(\set{x \in\mathbb{Z} : n\mid x}\) must be in \(\set{x \in\mathbb{Z} : mn\mid x}\) as well. In other words, if a number is a multiple of \(m\) and also is a multiple of \(n\text{,}\) then it is also a multiple of \(mn\text{.}\) This statement doesn’t really sound right since \(m\) can itself be a multiple of \(n\text{,}\) and thus, the hypothesis wouldn’t give us any more information other than the number being a multiple of \(m\) (since it automatically becomes a multiple of \(n\) in that situation). This also suggests that we find a counterexample to this statement. Moreover, this suggest what kind of a counterexample would work. Let’s see how we can create a counterexample.

8.6.6.

Scratchwork.
We see that this is a `prove or disprove’ question. This means that we need to figure out whether it is true or not.
The statement claims that every element of \(\set{x \in\mathbb{Z} : mn\mid x}\) is an element of \(\set{x \in\mathbb{Z} : m\mid x}\) and \(\set{x \in\mathbb{Z} : n\mid x}\text{.}\) In other words, every multiple of \(mn\) is a multiple of \(m\) and also a multiple of \(n\text{.}\) We can easily tell that this statement is true. So, we need to prove it.
Of course, we should be extra careful around statements that we can “easily” decide are true. Sometimes they are very deceptive. As a general rule, we should avoid using words like “easily” when we do mathematics.

8.6.9.

Scratchwork.
We see that since we are dealing with power sets in this question, we may not be able to use Venn diagrams to determine the set equality. This means that to see whether we want to prove or disprove the statement, we need to understand what the statement claims.
We can see that the statement says: “intersection of the power sets of \(A\) and \(B\) is the power set of their intersection”. In other words, it says that every set that is both a subset of \(A\) and \(B\) must be a subset of their intersection AND a subset of their intersection must be a subset of both \(A\) and \(B\text{.}\) This makes more sense. We can tell now that this statement is true. So, we can try to prove it.

8.6.10.

Scratchwork.
We can see that the statement says: “the union of the power sets of \(A\) and \(B\) is the power set of their union”. In other words, it claims that every set that is a subset of \(A\) or \(B\) must be a subset of their union and every subset of the union must be a subset of \(A\) or \(B\text{.}\) We can tell that the first part of the statement is indeed true, that is, every set that is a subset of \(A\) or \(B\) must be a subset of their union. But, second part of the statement doesn’t sound right, since the union of two sets can be “bigger” than both sets. This would create a problem since we can simply take the union of these two sets as the “subset” of the union, i.e. the element of the power set of the union. This suggests that we should be able to create a counterexample as long as \(A\cup B\neq A\) and \(A\cup B\neq B\text{.}\) Let’s see how we can create a counterexample.

8.6.11.

Scratchwork.
Let \(A\) be a set. We can check the statement for some numbers to see whether it works for, at least, sets with small sizes. This may also give us some ideas as to how to prove it for any size set.
We see that if \(A=\emptyset\text{,}\) then \(\pow{A}=\set{\emptyset}\text{.}\) This means that \(|\pow{A}|=1=2^0\text{.}\)
Now, assume that \(|A|=1\text{,}\) e.g. \(A=\set{1}\text{.}\) Then, we see that \(\pow{A}=\set{\emptyset, \set{1}}\text{.}\) Thus, \(|\pow{A}|=2=2^1\text{.}\)
Similarly, assume \(|A|=2\text{,}\) e.g. \(A=\set{1,2}\text{.}\) Then, we see that \(\pow{A}=\set{\emptyset, \set{1}, \set{2}, \set{1,2}}\text{.}\) Thus, \(|\pow{A}|=4=2^2\text{.}\)
Finally, if \(|A|=3\text{,}\) e.g. \(A=\set{1,2,3}\text{.}\) Then, we see that
\(\pow{A}=\set{\emptyset, \set{1}, \set{2}, \set{3}, \set {1,2},\set{1,3}, \set{2,3}, \set{1,2,3}}\text{.}\) Thus, \(|\pow{A}|=8=2^3\text{.}\)
This suggests that the stamement may, indeed, be true, and since it is a statement on the size of \(A\text{,}\) induction may be a useful tool.
To see how induction may work, we may need to understand how we can count the size of the power set of a set of size \(n+1\text{,}\) when we know the size for a set of size \(n\text{.}\) For that, we observe that
\begin{align*} \pow{\set{1,2,3}} \amp =\pow{\set{1,2}}\cup \set{\set{3},\set{1,3}, \set{2,3}, \set{1,2,3}}\\ \amp =\pow{\set{1,2}}\cup \set{\emptyset\cup\set{3},\set{1}\cup\set{3}, \set{2}\cup\set{3}, \set{1,2}\cup\set{3}}. \end{align*}
This suggests that when we add an element to a set, we double its power set’s size, since every subset that contains \(3\) is a subset, \(B\text{,}\) of \(\set{1,2}\) union \(\set{3}\text{.}\)

8.6.12.

Scratchwork.
Both statements are false. In the first case, notice that the empty set is an element of every power set. This means that the empty set is in the left-hand side, but not in the right-hand side.
The second statement is a little harder to disprove. Think about what happens when \(A\) and \(B\) have some, but not all, elements in common. For example, \(A = \set{1,2}\) and \(B = \set{2,3}\text{,}\) so that \(A-B = \set{1}\text{.}\) Then \(\set{1,2} \) is in the left-hand side, but not in the right-hand side. We can simplify this counter-example a little further, since we don’t actually need “\(3\)” to make it work.

8.6.13.

Scratchwork.
  1. In order to show that the two sets are equal, we need to show that each is a subset of the other. First, we need to show that
    \begin{equation*} x\in \bigcup_{n=3}^\infty \left(\frac{1}{n},1-\frac{1}{n}\right) \implies x\in (0,1). \end{equation*}
    This direction is more straightforward to prove, as \((1/n,1-1/n)\subset (0,1)\) for all \(n\in\mathbb{N}\text{,}\) \(n\geq 3\text{.}\) Then, we need to show that
    \begin{equation*} y\in (0,1) \implies y\in \bigcup_{n=3}^\infty \left(\frac{1}{n},1-\frac{1}{n}\right). \end{equation*}
    This direction is trickier. For \(y\in (0,1)\text{,}\) we need to find some \(N\in\mathbb{N}\) so that
    \begin{equation*} \frac{1}{N}\leq y \leq 1 - \frac{1}{N}. \end{equation*}
    The inequality \(1/N\leq y\) implies that we need \(N\geq 1/y\text{.}\) The inequality \(y\leq 1-1/N\) implies that we need \(N\geq 1/(1-y)\text{.}\) So we need to take \(N\) bigger than both \(1/y\) and \(1/(1-y)\text{;}\) for example, we can set \(N = \max\left\{\left\lceil\frac{1}{y} \right\rceil, \left\lceil \frac{1}{1-y} \right\rceil \right\}+1\)
    With this scratchwork, we can write up the proof.
  2. Similarly as the previous part, we need to show each of the sets is a subset of the other. First, we need to show that
    \begin{equation*} x\in [0,1] \implies x\in \bigcap_{n=1}^\infty \left(-\frac{1}{n}, 1+\frac{1}{n}\right) \end{equation*}
    This direction is fairly straightforward, since \([0,1]\subseteq (-1/n,1+1/n)\) for all \(n\in\mathbb{N}\text{.}\) The trickier part is showing that
    \begin{equation*} y \in \bigcap_{n=1}^\infty \left(-\frac{1}{n}, 1+\frac{1}{n}\right) \implies y\in [0,1]. \end{equation*}
    We can prove this by contrapositive, that is, instead prove
    \begin{equation*} y\not\in[0,1]\implies y\not\in\bigcap_{n=1}^\infty \left(-\frac{1}{n}, 1+\frac{1}{n}\right). \end{equation*}
    If \(y\not\in[0,1]\text{,}\) then either \(y \lt 0\) or \(y \gt 1\text{.}\) If \(y \lt 0\text{,}\) then we could find some \(N\in\mathbb{N}\) so that \(y \lt -1/N\text{.}\) But then \(y\not\in (-1/N,1+1/N)\text{,}\) which means that \(y\not\in \cap_{n=1}^\infty(-1/n,1+1/n)\text{.}\) If \(y \gt 1\text{,}\) then we could find some \(N\in\mathbb{N}\) so that \(1+1/N \lt y\text{.}\) Indeed, by rearranging, we see that this inequality will hold as long as \(N\gt 1/(y-1)\text{.}\) From here, we can similarly show the desired result.
    Now we can write the proof.

8.6.15.

Scratchwork.
  1. Any real number that is at least \(3\) is larger than or equal to every element of the set \([1,3]\text{.}\) That means that any real number that is at least \(3\) is an upper bound of \([1,3]\text{.}\) Since \(3\) is an element of \([1,3]\text{,}\) and any number less than \(3\) is not an upper bound for \([1,3]\text{,}\) we see that \(3\) is the least upper bound and the maximum of \([1,3]\text{.}\)
  2. We claim that \((1,3)\) has no maximum. In order to show this, we need to show that any \(x\in (1,3)\) is not an upper bound of the set.
    We claim that \(3\) is the supremum of \((1,3)\text{.}\) For any \(x\in (1,3)\text{,}\) \(x \lt 3\text{,}\) and therefore \(3\) is an upper bound of \((1,3)\text{.}\) Now we need to show that it is the least upper bound. That is, we need to show that if \(a\) is an upper bound of \((1,3)\text{,}\) then \(3\leq a\text{.}\) We will instead prove the converse of this statement: if \(a \lt 3\text{,}\) then \(a\) is not an upper bound of \((1,3)\text{.}\)
  3. It is important to realise that this is a subset of integers, not reals. Then we can simplify the set before we determine its maximum or supremum. The inequality \(|2(m-4)|\leq15\) holds if and only if
    \begin{equation*} -15 \leq 2(m-4)\leq 15. \end{equation*}
    Dividing by 2, we then have
    \begin{equation*} -\frac{15}{2} \leq m-4\leq \frac{15}{2} \end{equation*}
    and then adding \(4\text{,}\)
    \begin{equation*} -\frac{7}{2} \leq m\leq \frac{23}{2}. \end{equation*}
    Since each of these steps can be reversed, we see that this inequality is equivalent to the original one. Hence
    \begin{equation*} \set{m\in\mathbb{Z}: |2(m-4)|\leq 15}=\set{m\in\mathbb{Z}:- \frac{7}{2}\leq m\leq \frac{23}{2} }. \end{equation*}
    In this form, it’s easier to see that the maximum and supremum of the set is \(11\text{:}\) indeed, any integer that is at most \(\frac{23}{2} = 11 + \frac{1}{2}\) must be at most \(11\text{,}\) so \(11\) is an upper bound for the set. Since \(11\) is an element of the set, it is the maximum. Since any element less than \(11\) cannot be an upper bound for the set, we see that \(11\) is also the supremum.
  4. Let
    \begin{equation*} S=\set{2-\frac{1}{n}:n\in\mathbb{N}}. \end{equation*}
    We claim that the set has no maximum. We need to show that for any \(x\in S\text{,}\) there is some other element of \(S\) that is greater than \(x\text{;}\) then \(x\) will not be an upper bound for \(S\text{,}\) and hence not its maximum.
    We claim that the supremum of \(S\) is \(2\text{.}\) First, we’ll need to show that \(2\) is an upper bound for \(S\text{.}\) Then, we’ll need to show that for any \(a \lt 2\text{,}\) \(a\) is not an upper bound of \(S\text{.}\) In order to do that, we need to find some \(n\in\mathbb{N}\) so that \(a \lt 2-1/n\text{.}\) Rearranging this, we have \(1/n \lt 2-a\text{.}\) Since \(a \lt 2\text{,}\) we have \(2-a \gt 0\text{,}\) and so the inequality \(1/n \lt 2-a\) may be rearranged to \(1/(2-a) \lt n\text{.}\) So the choice of \(n=\lceil 1/(2-a)\rceil+1\) will work.
  5. Recall that \(\cos(\theta)=1\) holds if and only if \(\theta =2\pi k\) for some \(k\in\mathbb{Z}\text{.}\) Hence, if \(\cos(2x)=1\text{,}\) we must have \(x=\pi k\) for some \(k\in\mathbb{Z}\text{.}\) So
    \begin{equation*} \set{x\in \mathbb{R}:\cos(2x)=1}=\set{\pi k:k\in\mathbb{Z}}. \end{equation*}
    We claim that this set has no upper bound; this means that the set cannot have a maximum or a supremum, since both the maximum and supremum would be upper bounds of the set.

8.6.16.

Scratchwork.
  1. We need to show that \(a=\max\{s,t\}\) satisfies the following two properties:
    • \(a\) is an upper bound for \(S\cup T\text{,}\) so \(x\leq a\) for all \(x\in S\cup T\text{,}\) and
    • \(a\leq b\) for any upper bound \(b\) of \(S\cup T\text{.}\)
  2. Let’s look at some examples for \(S\) and \(T\) to see if we can find any patterns.
    • Say \(S=(0,2)\) and \(T=(0,1)\text{,}\) so that \(s=2\) and \(t=1\text{.}\) Then \(S\cap T=T\text{,}\) and so \(\sup(S\cap T)=1=t\text{.}\)
    • Suppose \(S=(0,2)\) and \(T=(1,3)\text{,}\) so that \(s=2\) and \(t=3\text{.}\) Then \(S\cap T=(1,2)\text{,}\) and so \(\sup(S\cap T)=2=s\text{.}\)
    From these examples, we may believe that \(\sup(S\cap T)=\min(S,T)\text{.}\) However, we haven’t yet looked at an example where \(S\) and \(T\) are disjoint.
    • Say \(S=(0,1)\) and \(T=(1,3)\text{,}\) in which case \(s=1\) and \(t=3\text{.}\) Then \(S\cap T=\emptyset\text{.}\) What’s the least upper bound of the empty set? We claim that any real number is an upper bound for the empty set. Indeed, suppose \(a\in \mathbb{R}\text{.}\) The statement, if \(s\in\emptyset\) then \(s\leq a\text{,}\) is true, since the hypothesis is always false! Since every real number of \(\emptyset\) is an upper bound, it can’t have a least upper bound.
  3. We need to show that \(a=s+t\) satisfies the following two properties:
    • \(a\) is an upper bound for \(S+T\text{.}\)
    • If \(b\) is an upper bound of \(S+T\text{,}\) then \(a\leq b\text{.}\)
    The second part is trickier to prove, and we’ll try to prove the contrapositive instead:
    If \(b \lt a\text{,}\) then \(b\) is not an upper bound of \(S+T\text{.}\)
    This means that we need to find some \(x\in S+T\) such that \(x \gt b\text{.}\) We want to write \(b\) as a sum of two elements, say \(u\) and \(v\text{,}\) that are less than \(s\) and \(t\) respectively. If we have such an \(u\) and \(v\text{,}\) then they will not be upper bounds for \(S\) and \(T\text{,}\) respectively. Then, by definition, there will be \(c\in S\) and \(d\in T\) so that \(c \gt u\) and \(d \gt v\text{.}\) But then \(c+d \gt u+v=b\text{,}\) so \(b\) would not be an upper bound for \(S+T\text{.}\)
    How can we find these elements \(u\) and \(v\text{?}\) There are many choices, but we’ll set
    \begin{equation*} u = s-\epsilon \qquad \text{and} \qquad v = t-\epsilon \end{equation*}
    where \(\epsilon = \frac{a-b}{2}\text{.}\) Then
    \begin{equation*} u+v = s+t-2\epsilon = a-(b-a) = b. \end{equation*}

8.6.17.

Scratchwork.
By definition, we need to show that given any \(\epsilon\gt 0\text{,}\) there is some \(N\in \mathbb{N}\) such that for all \(n\geq N\text{,}\) \(|a_n-a|\lt \epsilon\text{.}\) Unravelling the inequality, we need to show
\begin{equation*} -\epsilon \lt a_n-a\lt \epsilon\iff a-\epsilon \lt a_n \lt a+\epsilon. \end{equation*}
Since \(a\) is the supremum of \(\{a_n:n\in\mathbb{N}\}\text{,}\) we know that \(a_n\leq a\) for all \(n\in\mathbb{N}\text{.}\) So the inequality \(a_n\lt a+\epsilon\) is satisfied for all \(n\in\mathbb{N}\text{.}\) Therefore, we can focus on satisfying the inequality
\begin{equation*} a-\epsilon \lt a_n. \end{equation*}
To establish this inequality for sufficiently large \(n\text{,}\) we need to use the fact that \(a\) is the least upper bound of \(\{a_n:n\in\mathbb{N}\}\text{.}\) Since \(\epsilon \gt 0\text{,}\) we know that \(a-\epsilon \lt a\text{.}\) Since \(a-\epsilon\lt a\) and \(a\) is the least upper bound, we know that \(a-\epsilon\) is not an upper bound of the set \(\{a_n:n\in\mathbb{N}\}\text{.}\) Therefore, there is some \(N\in\mathbb{N}\) such that \(a-\epsilon \lt a_N\text{.}\) Finally, we just need to use the fact that \(a_n\) is increasing to show that \(a-\epsilon\lt a_n\) for all \(n\geq N\text{.}\)

9 Relations
9.7 Exercises

9.7.4.

Scratchwork.
  1. For the first relation, we see that the definition of the relation is symmetric with respect to \(f\) and \(g\text{.}\) This suggests that the relation is symmetric as well. Also, we can see that it is reflexive, since a function is equal to itself at every point. However we see that it may not be transitive since for two functions to be related, they have to be same at a point, and if we have \(f\rel g\) and \(g\rel h\text{,}\) we may have two different points \(x \neq y\in\mathbb R\) such that \(f(x)=g(x)\) and \(g(y)=h(y)\text{,}\) but this does not mean that \(f\) and \(h\) are equal anywhere. But of course, we will need to find a counterexample to show that the relation is not transitive when we are writing the proof. Simply saying that we can find \(f,g, h\) and \(x,y\in\mathbb R\) satisfying the condition above is not enough.
  2. For the second relation, we see that it is symmetric since the definition of the relation is symmetric with respect to \(x\) and \(y\text{.}\) But, we can also see that is is not reflexive since \((1,1)\notin\ \rel[S]\) and also not transitive since every integer is related to \(0\text{,}\) that is, we can take \((1,0)\in\rel[S]\text{,}\) \((0,1)\in\rel[S]\text{,}\) but we see that \((1,1)\notin\rel[S]\) as mentioned.

9.7.7.

Scratchwork.
The hint suggests that this relation says for the nonempty set \(E\text{,}\) \(q\in E\text{,}\) and \(A,B\in\mathcal P(E)\text{,}\) \(A\rel B\) if \(q\) is in both of them or in neither of them. This observation immediately suggests that the relation is reflexive and symmetric. Moreover, we see that if \(A\rel B\) and \(B\rel C\text{,}\) then we see that \(q\) is in all three sets \(A, B\text{,}\) and \(C\text{;}\) or in none of them. In either case, we see that \(A\rel C\text{.}\) Thus, the relation should be transitive too.

9.7.13.

Scratchwork.
For the first part, we first realize the relation \(R = \{(a,b) \in \mathbb{Z} \times \mathbb{Z} : n\mid (a-b)\}\) is nothing but the congruence modulo \(n\) relation. Moreover, we see that if \(R\) is not sparse than there are \(a,b\in\mathbb Z\) such that \(a\equiv b \pmod n\) and either \(a\equiv (b+1) \pmod n\) or \((a+1)\equiv b \pmod n\text{.}\) Since these cases are symmetric with respect to \(a\) and \(b\text{,}\) WLOG, we can assume that \(a\equiv (b+1) \pmod n\text{.}\) Then, we see that \(n\mid (b-a)\) and \(n\mid ((b+1)-a)\text{,}\) that is \(n\mid ((b-a)+1)\text{.}\) Therefore, as the hint suggested, we get \(n=1\) (since \(n\in\mathbb N\)). We can also see that if \(n=1\text{,}\) then the congruence relation cannot be sparse, since every element will be related to every nonzero element.
This final observation is also going to be useful in the second part, since we know that the relation \(R = \{(a,b) \in \mathbb{Z} \times \mathbb{Z} : 1\mid (a-b)\}=\mathbb Z\times \mathbb Z\) is an equivalence relation, and it is not sparse.
Now, we can write this nicely as a proof.

9.7.17.

Scratchwork.
  1. We want to show that \(nk\equiv1\mod{p}\text{.}\) Then by definition, we want to show that \(p\mid nk-1\text{,}\) that is, there’s some \(\ell\in\mathbb{Z}\) such that \(p\ell=nk-1\text{.}\) Let’s rewrite this equation as \(1=kn+(-\ell)p\text{.}\) This equation may look familiar from Bézout’s identity, which tells us that if \(a,b\in \mathbb{Z}\text{,}\) with at least one of them non-zero, then there are some \(x,y\in\mathbb{Z}\) so that \(ax+by=\gcd(a,b)\text{.}\) In order to show that there is some \(\ell\in\mathbb{Z}\) so that \(1=kn+(-\ell)p\text{,}\) it suffices, by Bézout’s identity, to show that \(\gcd(n,p)=1\text{.}\)
    To show that \(\gcd(n,p)=1\text{,}\) we’ll use the hypothesis that \(n\not\equiv0\mod{p}\text{.}\) By definition, this means that \(p\nmid (n-0)\text{,}\) so \(p\nmid n\text{.}\) Since \(p\) is prime, its only divisors are \(1\) and \(p\text{.}\) Since \(p\) isn’t a divisor of \(n\text{,}\) we must have \(\gcd(n,p)=1\text{.}\)
    Now we can write the proof.
  2. Let’s try to find such an \(n\) when \(p\) is not prime; we’ll use \(p=4\) since it is a small composite number. We need to choose \(n\) so that \(\gcd(n,p)\neq1\text{;}\) if \(n\) is such that \(\gcd(n,p)=1\text{,}\) then we could use Bézout’s identity as in part (a) to find some \(k\) with \(nk\equiv1\mod{p}\text{.}\) So let’s try \(n=2\text{.}\) We need to check that \(nk\not\equiv1\pmod{4}\) for each \(k\in\set{0,1,2,3}\text{,}\) since these represent all equivalence classes mod \(4\text{.}\) We have
    \begin{equation*} 2\cdot0\equiv0\pmod{4}, \quad 2\cdot1\equiv2\pmod{4}, \quad 2\cdot2\equiv0\pmod{4}, \quad 2\cdot3\equiv2\pmod{4}, \end{equation*}
    so there is no \(k\in\mathbb{Z}\) with \(2k\equiv1\mod{4}\text{.}\)
    We chose \(n=2\) because then \(nk\) is always an even number, however in order to have \(nk \equiv 1 \mod{4}\text{,}\) we need \(nk\) to be odd. Since a number cannot be even and odd at the same time, that cannot happen.

9.7.18.

Scratchwork.
Lemma 9.5.9 proves the statement in the special case that \(\gcd(a,d)=1\text{.}\) We’ll try to modify the proof of that lemma. Bézout’s identity tells us that there are \(x,y\in\mathbb Z\) so that
\begin{equation*} \gcd(a,d)=xa+yd. \end{equation*}
We can rewrite this as
\begin{equation*} 1 = x\cdot \frac{a}{\gcd(a,d)}+y\cdot\frac{d}{\gcd(a,d)} \end{equation*}
since \(\gcd(a,d)\) divides both \(a\) and \(d\text{.}\) Now multiplying the entire equation by \(b\text{,}\) we have
\begin{equation*} b= x\cdot \frac{ab}{\gcd(a,d)}+yb\cdot\frac{d}{\gcd(a,d)}. \end{equation*}
The last step is to use that \(d\mid ab\) in order to show that this equation implies that \(d/\gcd(a,d)\) divides \(b\text{.}\)

9.7.19.

Scratchwork.
  1. Bézout’s identity tells us that there are \(x,y\in\mathbb Z\) so that \(\gcd(a,b)=ax+by\text{.}\) This equation implies that any divisor of both \(a\) and \(b\) must divide \(\gcd(a,b)\text{.}\)
  2. We will try to show that \(m\gcd(a,b)\leq \gcd(ma,mb)\) and \(\gcd(ma,mb)\leq m\gcd(a,b)\text{.}\)
    We claim that \(m\gcd(a,b)\mid \gcd(ma,mb)\text{.}\) We know that \(d=\gcd(a,b)\) divides both \(a\) and \(b\text{;}\) from this, we will be able to show that \(md\) divides both \(ma\) and \(mb\text{.}\)Then \(md\leq \gcd(a,b)\text{.}\)
    Now let’s consider the converse, \(\gcd(ma,mb)\leq m\gcd(a,b)\text{.}\) Now \(m\) divides both \(ma\) and \(mb\text{,}\) so by part (a), \(m\mid \gcd(ma,mb)\text{.}\) Then there is some \(e\in\mathbb Z\) so that \(me=\gcd(ma,mb)\text{.}\) To prove that \(\gcd(ma,mb)\leq m\gcd(a,b)\text{,}\) we can show that \(e\leq\gcd(a,b)\text{.}\) Let’s try to prove that \(e\) divides both \(a\) and \(b\text{.}\) By definition of \(\gcd(ma,mb)\) there are \(u,v\in\mathbb Z\) so that \(meu=\gcd(ma,mb)u=ma\) and \(mev=\gcd(ma,mb)v=ma\text{;}\) then we can cancel the factors of \(m\) to get that \(e\mid a\) and \(e\mid b\text{.}\)
  3. Take \(a=b=c=2\text{.}\) Then \(gcd(a,b)=\gcd(c,b)=\gcd(ac,b)=2\text{,}\) and so \(\gcd(ac,b)\neq \gcd(a,b)\cdot\gcd(c,b)\text{.}\)

10 Functions
10.8 Exercises

10.8.2.

Scratchwork.
Let us start by rearranging things a little. We must have \(ax+by=6\) which means that
\begin{equation*} y = \frac{6-ax}{b}. \end{equation*}
So for any given \(x\) there is only a single value of \(y\text{.}\)
But we also require that \(y \in \mathbb{Z}\text{,}\) which means that we must have
\begin{equation*} b \mid (6-ax) \end{equation*}
for every \(x \in \mathbb{Z}\text{.}\) Setting \(x=0\) immediately tells us that \(b \mid 6\text{,}\) and so \(b \in \set{1,2,3,6}\text{.}\)
  • When \(b=1\text{,}\) then \(y=6-ax \in \mathbb{Z}\) for any \(a \in \mathbb{N}\text{.}\)
  • When \(b=2\text{,}\) then \(y=\frac{6-ax}{2} = 3-\frac{ax}{2}\) will be an integer for all \(x\) provided \(2 \mid a\text{.}\)
  • When \(b=3\text{,}\) then \(y=\frac{6-ax}{3} = 2-\frac{ax}{3}\) will be an integer for all \(x\) provided \(3 \mid a\text{.}\)
  • And finally, when \(b=3\text{,}\) then \(y=\frac{6-ax}{6} = 1-\frac{ax}{6}\) will be an integer for all \(x\) provided \(6 \mid a\text{.}\)
That is, \(\phi\) is a function provided \(b \in \set{1,2,3,6}\) (that is \(b \mid 6\)) and \(b \mid a\text{.}\)
Now, strictly speaking we have shown that if \(\phi\) is a function, then we have \(b \mid 6, b\mid a\text{.}\) We can also readily show that when \(b \mid 6, b \mid a\) then \(\phi\) is a function.
So now assume that \(b \mid 6, b \mid a\text{,}\) and sot \(6=kb, a=\ell b\) for integers \(k,\ell\text{.}\) Hence
\begin{equation*} y = \frac{6-ax}{b} = \frac{kb - \ell b x}{b} = k-\ell x \end{equation*}
and thus if \(x \in \mathbb{Z}\) then \(y \in \mathbb{Z}\text{,}\) and so we have a function.

10.8.3.

Scratchwork.
As part of our proof we need to prove that \(f(x) \in [-1,1]\text{.}\) To do this let us start with the inequality we wish to prove, namely
\begin{gather*} -1 \leq \frac{2x}{1+x^2} \leq 1 \end{gather*}
Now since \(x \in \mathbb{R}\) we know that \(x^2 \geq 0\) and so \(1+x^2 \geq 1\text{.}\) Hence, we have
\begin{gather*} -1-x^2 \leq 2x \leq 1+x^2 \end{gather*}
Subtracting \(2x\) from everything gives
\begin{align*} -1-2x-x^2 \amp \leq 0 \leq 1-2x+x^2 \\ -(1+x)^2 \amp \leq 0 \leq (1-x)^2 \end{align*}
which we know to be true. In our proof we will use this chain of reasoning, but in reverse; starting from what we know is true and going back to the inequality we need.
We will also need to show that given \(y \in [-1,1]\) that \(y \in f(\mathbb{R})\text{.}\) That is, there is some \(x \in \mathbb{R}\) so that \(f(x)=y\text{.}\) Our proof doesn’t have to explain how we found that \(x\)-value, just that it works. Let us find it now:
\begin{align*} y \amp = f(x) = \frac{2x}{1+x^2} \\ yx^2 -2x +y \amp = 0 \\ x \amp = \frac{2 \pm \sqrt{4-4y^2}}{2y} = \frac{2 \pm 2\sqrt{1-y^2}}{2y} \end{align*}
This looks like it should work since \(-1 \leq y \leq 1\) and so \(0 \leq y^2 \leq 1\text{,}\) but we might have a problem when \(y=0\text{.}\) However, when \(y=0\) we can just take \(x=0\text{.}\)

10.8.6.

Scratchwork.
We see that for any values of \(a\) and \(b\) this function \(f(x) = x^2+ax+b\) is a quadratic function. This means that we wouldn’t expect the function to be injective or surjective. Moreover, we see that we can complete the square and get \(f(x)=\left(x+\frac{a}{2}\right)^2+\left(b-\frac{a^2}{4}\right)\text{.}\) This tells us the minimum of the function (which we can use to show that the function is not surjective) and also the vertex of the graph of the function (which we can use to show that the function is not injective).

10.8.7.

Scratchwork.
We see that the condition in the question tells us that \(f(n)\leq n\) for all \(n\in\mathbb N\) and that \(f\) is an injective function. If we want to understand the function, we see that for large \(n\in\mathbb N\text{,}\) say \(n=1000000\text{,}\) we have \(f(1000000)\leq 1000000\text{.}\) This tells us that \(f(1000000)\) has \(1000000\) different options. If we also want to consider that \(f\) is an injective function, it may not be clear as to which of those options \(f(1000000)\) can take.
Instead, we can try to understand what happens to \(f\) when \(n\in\mathbb N\) is small. This will give us fewer options for \(f(n)\) and maybe we can even figure out some values of \(f\text{.}\)
Now, if we check \(n=1\text{,}\) we see \(f(1)\leq 1\text{.}\) Since \(f:\mathbb N \rightarrow \mathbb N\text{,}\) this tells us that \(f(1)=1\text{.}\) Similarly, if we check \(n=2\text{,}\) we see that \(f(2)\leq 2\text{.}\) This means that \(f(2)=1\) or \(f(2)=2\text{.}\) But, since \(f\) is injective, we see that \(f(2)\neq 1\text{.}\) Thus \(f(2)=2\text{.}\)
This suggests that if we keep going like this, we will end up with \(f(n)=n\text{.}\) But, we also realize that if we want to “keep going like this”, we need to apply induction. Moreover, since we want to know something about the values of the function at any point to be able to determine the next one since we want to make sure that the function is injective, we may need strong induction.

10.8.8.

Scratchwork.
To prove that the function is injective we start by assuming \(f(a)=f(b)\) and then prove that this implies that \(a=b\text{.}\) This requires a little juggling and factoring of polynomials. To prove that it is surjective we take some \(y \in [5,\infty)\) and then show that there is an \(x \in [3,\infty)\) so that \(y=f(x)\text{.}\) Now, the proof doesn’t have to show how we came up with that \(x\text{,}\) only that it works. We find it by solving \(y=f(x)\text{:}\)
\begin{align*} x^2-6x+16 \amp =y \amp \text{complete the square} (x^2-6x+9)+5 \amp =y\\ (x-3)^2 \amp = y-5 \amp \text{careful with square-roots}\\ (x-3) \amp = \pm \sqrt{y-5}\\ x \amp = 3 \pm \sqrt{y-5} \end{align*}
Now since \(y \geq 5\text{,}\) the square root is defined. And since we want \(x \geq 3\text{,}\) we take the positive branch. That is, given a \(y\)-value in the codomain, set \(x = 3+\sqrt{y-5}\text{.}\)

10.8.15.

Scratchwork.
When we look at the relation defined in the question, we can easily see that it is reflexive (since equality is reflexive), and symmetric (since equality is symmetric). Moreover, if \((x,y)\rel (s,t)\) and \((s,t)\rel (a,b)\text{,}\) we see that \(x^2+y^2=s^2+t^2=a^2+b^2\text{.}\) Therefore it also is a transitive relation. As we can see, trying to show that this relation is an equivalence relation is not too hard to show.
Now, let’s try to understand the equivalence classes of this relation. Let \((s,t)\in\mathbb R^2\text{,}\) then we see
\begin{align*} [(s,t)] \amp =\set{(x,y)\in\mathbb R^2\colon (s,t)\rel (x,y)}\\ \amp =\set{(x,y)\in\mathbb R^2\colon x^2+y^2=s^2+t^2}\\ \amp =\set{(x,y)\in\mathbb R^2\colon \sqrt{x^2+y^2}=\sqrt{s^2+t^2}}. \end{align*}
This means that \([(s,t)]\) is the set of all points on the plane that have the same distance from the origin, that is, the points on the circle, centred at origin of radius \(\sqrt{s^2+t^2}\text{.}\)
This is an important observation in understanding whether \(f\) is a function. We know that for \(f\) to be a function, we need to show that
  • \(f\) is defined on all its domain — this is easy since for any \(x,y\) we know that \(x^2+y^2 \geq 0\) and so \(\sqrt{x^2+y^2} \in \mathbb{R}\) as required.
  • if we input different representatives of the same equivalence class, we get the same result, namely
    \begin{equation*} \Big([(x,y)]=[(s,t)]\Big)\implies f\big([(x,y)]\big)=f\big([(s,t)]\big). \end{equation*}
    We also know that if \([(x,y)]=[(s,t)]\text{,}\) then the points \((x,y)\) and \((s,t)\) are on the same circle centred at the origin. Then, since \(f\) takes an equivalence class (which is a circle) and sends it to the radius of the circle, we see that it shouldn’t matter which points on the circle we pick as the representative of the equivalence class, we get the same output for \(f\text{.}\)
This means that \(f\) defines a function.
Once we see that \(f\) is a function we can also easily see that it is surjective since for all \(z\in [0,\infty)\text{,}\) we can take \([(x,y)]=[(0,z)]\text{,}\) and see that
\begin{equation*} f\big([(x,y)]\big)=f\big([(0,z)]\big)=\sqrt{z^2}=z. \end{equation*}
To show the injectivity we remember, again, that \(f\) takes the circles centred at the origin and sends them to their radius. This suggests that if
\begin{equation*} f\big([(x,y)]\big)=f\big([(s,t)]\big), \end{equation*}
then the radius of the circle defined by \([(x,y)]\) and \([(s,t)]\) must be the same. Since these circles are centred at the origin, we see that they must be the same, that is,
\begin{equation*} [(x,y)]=[(s,t)], \end{equation*}
which means that \(f\) is injective.

10.8.16.

Scratchwork.
  1. Let \([z]_n \in \mathbb{Z}_n\text{.}\) We need to show that \(f([z]_n) \in \set{0,1,\dots,n-1}\text{.}\) We know that \(z \in [z]_n\text{,}\) and then by Euclidean division, we know that there exist \(q,r \in \mathbb{Z}\) so that \(z = qn +r\) and \(r \in \set{0,1,\dots,n-1}\text{.}\) Hence \(f([z]_n) = r \in \set{0,1,\dots,n-1}\) as required.
    Now suppose that \([x]_n=[y]_n\text{.}\) We need to show that \(f([x]_n)=f([y]_n)\text{.}\) Untangling the definitions, we know that \(n\) divides \(x-y\text{,}\) and we need to show that \(x\) and \(y\) have the same remainder upon division by \(n\text{.}\) By Euclidean division, we can write \(x=q_xn+r_x\) and \(y=q_yn+r_y\) for some integers \(q_x,q_y,r_x,r_y\) with \(0\leq r_x,r_y\leq n-1\text{.}\) Then \(x-y=(q_x-q_y)n+(r_x-r_y)\text{.}\) Since \(n\) divides \(x-y\) and \((q_x-q_y)n\text{,}\) we can show from this equation that \(n\) divides \(r_x-r_y\text{.}\) Now we can use the restrictions on the size of the remainders, namely that \(0\leq r_x,r_y\leq n-1\text{,}\) combined with the fact that \(n\mid r_x-r_y\) to show that \(r_x-r_y=0\text{.}\)
  2. In order to show that \(f\) is a bijection, we need to show that \(f\) is both injective and surjective.
    For injectivity, we assume that \(f([x]_n)=f([y]_n)\) for some \([x]_n,[y]_n\in\mathbb{Z}_n\text{,}\) and we need to show that \([x]_n=[y]_n\text{.}\) Note that \(f([x]_n)=f([y]_n)\) implies that \(x\) and \(y\) have the same remainder when divided by \(n\text{.}\) From this observation, we can show that \(n\) divides \(x-y\text{,}\) and consequently, \(x\) and \(y\) are congruent modulo \(n\text{.}\)
    For surjectivity, we need to show that for any \(r\in \set{0,1,\dots, n-1}\text{,}\) the codomain of the function, there is some \([x]_n\in\mathbb{Z}_n\) so that \(f([x]_n)=r\text{.}\) This is the same as saying that there is some \(x\in\mathbb{Z}\) so that its remainder upon division by \(n\) is \(r\text{.}\)

10.8.20.

Scratchwork.
We do some rough computation here to try to figure out which statements are true or false. When we write up the proof, we will include more detail in these computations.
  1. Fix \(x\in \mathbb{R}\text{.}\) We compute, starting with the left hand side. The computation below uses the definition of function composition, followed by the definition of function addition.
    \begin{align*} (f\circ (g+h))(x) \amp = f((g+h)(x))\\ \amp = f(g(x)+h(x)) \end{align*}
    Now, for the right hand side, we compute using the definitions of function addition and composition
    \begin{align*} (f\circ g + f \circ h)(x) \amp = (f\circ g)(x) + (f\circ h)(x)\\ \amp =f(g(x)) + f(h(x)) \end{align*}
    Most functions \(f\) do not have the property that
    \begin{equation*} f(g(x)+h(x)) = f(g(x))+f(h(x)). \end{equation*}
    This is a property of linear functions. Therefore, we can create a counterexample using a nonlinear function such as \(f(x)=x^2\text{.}\)
  2. Fix \(x\in \mathbb{R}\text{.}\) We compute, starting from the left hand side. By the definition of function composition,
    \begin{align*} ((g+h)\circ f)(x) \amp = (g+h)(f(x)) \amp \text{addition} \\ \amp = g(f(x))+h(f(x)) \amp \text{composition}\\ \amp =(g\circ f)(x) + (h\circ f)(x) \amp \text{addition}\\ \amp = \left( g\circ f + h\circ f \right)(x) \end{align*}
    This computation did not rely on our choice of \(x\text{,}\) so it holds for all \(x\in \mathbb{R}\text{.}\) Thus, we can show
    \begin{equation*} (g+h)\circ f = g\circ f + h\circ f. \end{equation*}
  3. We first use the definition of function composition to see that
    \begin{align*} \left(\frac{1}{f}\circ g \right)(x) \amp = \left(\frac{1}{f}\right)(g(x)). \end{align*}
    By the definition of function division, we have
    \begin{align*} \amp = \frac{1}{f(g(x))} \end{align*}
    By the definition of function composition, we have
    \begin{align*} \amp =\frac{1}{(f\circ g)(x)} \end{align*}
    Using the definition of function division yields
    \begin{align*} \amp =\left(\frac{1}{f\circ g}\right)(x) \end{align*}
    This computation did not rely on our choice of \(x\text{,}\) so it holds for all \(x\in \mathbb{R}\text{.}\) Thus, we can show
    \begin{equation*} \left(\frac{1}{f}\right)\circ g = \frac{1}{f\circ g}. \end{equation*}
  4. Fix \(x\in \mathbb{R}\text{.}\) We start with the left hand side, and use the definition of function division to compute,
    \begin{align*} \left(\frac{1}{f\circ g}\right)(x) \amp = \frac{1}{(f\circ g)(x)} \amp \text{composition}\\ \amp = \frac{1}{f(g(x))} \end{align*}
    On the other hand, we start from the right hand side. Using the definition of function composition to compute
    \begin{align*} \left(f\circ \frac{1}{g}\right)(x) \amp = f\left(\left(\frac{1}{g}\right)(x)\right) \amp \text{division}\\ \amp =f\left(\frac{1}{g(x)}\right) \end{align*}
    In order for
    \begin{equation*} \frac{1}{f(g(x))} = f\left(\frac{1}{g(x)}\right), \end{equation*}
    \(f\) must have the property that
    \begin{equation*} \frac{1}{f(x)} = f\left(\frac{1}{x}\right). \end{equation*}
    This is not true for every function \(f:\mathbb{R}\to \mathbb{R}\text{.}\) For example, if \(f(x)=x+2\text{,}\) we have
    \begin{equation*} \frac{1}{f(x)} = \frac{1}{x+2} \neq \frac{1}{x}+2 = f\left(\frac{1}{x}\right). \end{equation*}

10.8.21.

Scratchwork.
  1. We will try to construct a function \(f\) and sets \(X,W\) such that \(f(W\cap X)\) has few elements, but \(f(W)\) and \(f(X)\) have many elements in common. One way of accomplishing this is to make \(W\cap X\) small (for example, having only one element in common), but designing the function \(f\) to produce the same output for distinct elements of \(X\) and \(W\text{.}\)
    As a counterexample, take \(f:\set{-1,0,1}\rightarrow \set{0,1}\) defined as \(f(x)=|x|\) and consider the subsets \(W=\set{-1,0}\) and \(X\set{0,1}\text{.}\) Then we see \(f (W\cap X)=f(\set{0})=\set{0}\) and \(f(W)=f(X)=\set{0,1}\text{.}\) Hence \(f (W\cap X) =\set{0}\neq \set{0,1}= f (W)\cap f (X)\text{.}\)
  2. Recall from Theorem 10.3.6(ii), \(f(f^{-1}(Y))\subseteq Y\) for any \(Y\subseteq B\text{.}\) Therefore we need to find an example where \(Y\) is not a subset of \(f(f^{-1}(Y))\text{.}\) This means that we need to find a function that is not surjective, i.e. not every element of Y is in the range of the function.
    As a counterexample, we take \(f:\set{1,2}\rightarrow\set{1,2,3}\text{,}\) defined as \(f(x)=x\) and let \(Y=\set{2,3}\text{.}\) Notice that there is no element in the domain that maps to \(3\text{,}\) so \(f\) fails to be surjective.

10.8.25.

Scratchwork.
We have to show the function is both injective and surjective.
  • Injective: Assume that \(f(x)=f(z)\) and show that \(x=z\text{;}\) this requires some simple manipulations of polynomials.
  • Surjective: Given some \(y\) in the codomain we have to find an \(x\) in the domain so that \(f(x)=y\text{.}\) Our proof does not have to contain how we found that \(x\)-value, just that it is in the domain and that it satisfies \(f(x)=y\text{.}\) To find the \(x\) we solve
    \begin{align*} y \amp = \dfrac{x+1}{x+2}\\ (x+2)y \amp = x+1\\ xy-x \amp = 1-2y\\ x \amp = \frac{1-2y}{y-1} = -2 - \frac{1}{y-1} \end{align*}
    where we have done a little partial fractions at the last step.
    Notice that since \(y \neq 1\) we know that \(x \in \mathbb{R}\text{,}\) and since \(\frac{1}{y-1} \neq 0 \text{,}\) we know that \(x \neq -2\text{.}\)
This gives us the result we wanted. Now, we can write the proof.

10.8.26.

Scratchwork.
In this question, we can try to show that \(f\) is an injective and surjective function, and conclude that \(f\) has an inverse, and calculate the inverse afterwards.
However, if we know what the inverse of \(f\) should look like, call it a new function \(g\text{,}\) we can check whether \(g\) is a right and also a left inverse of \(f\text{.}\) If so, using Lemma 10.6.3 and Lemma 10.6.4, we can conclude that \(f\) is bijective and moreover \(g\) is its inverse.
We can find what the inverse of \(f\) should look like quite easily in this question. We see that \(f\) takes even numbers to odd numbers and the odd numbers to even numbers. For example, \(f(5)=5+7=12\text{.}\) This means that if we want to find, \(g\text{,}\) the inverse of \(f\) we should have \(g(12)=5=12-7\) (since we added \(7\) to \(5\) to get \(12\)), that is \(g(n)=n-7\) for \(n\) even. Similarly for \(n\) even, we see that \(f(n)=-n+3\text{,}\) so we must have \(g(-n+3)=n=-((-n+3)-3)\text{,}\) that is, \(g(n)=-n+3\) for \(n\) odd. Now, we can check whether this function \(g\) is the right- and the left-inverse of \(f\text{.}\)

11 Proof by contradiction
11.3 Exercises

11.3.1.

Scratchwork.
Proof by contradiction can be a useful tool for proving statements where we want to show the non-existence of an object satisfying certain conditions. This is because when we assume for a contradiction that such an object exists, then we would have some structure we can work with and, hopefully, get a contradiction with our prior assumptions or knowledge.

11.3.2.

Scratchwork.
We’ll try to prove the statement by contradiction, by assuming that both \(a\) and \(b\) are odd. Using the definition of \(a\) and \(b\) being odd, we can prove that \(a^2+b^2=4n+2\) for some \(n\in\mathbb Z\text{.}\) Since \(c^2=a^2+b^2\text{,}\) this implies that \(c^2\) is even, and so \(c\) is even. (The contrapositive of this statement is given in Result 3.2.10) But then \(4\) will divide \(c^2\text{.}\) This contradicts the equation \(c^2=a^2+b^2=4n+2\text{,}\) which implies that \(c^2\) is not divisible by \(4\text{.}\)

11.3.3.

Scratchwork.
In order to prove the statement by contradiction, we need to assume that there is some \(k\in\mathbb{Z}\) so that \(ak\equiv1\mod{n}\text{,}\) and then show that this leads to a contradiction, implying that the statement must be true. From the equation \(ak\equiv1\mod{n}\text{,}\) we know that \(n\mid (ak-1)\text{.}\) From here we can show that \(\gcd(a,n)\mid 1\text{.}\) This contradicts that \(\gcd(a,n)\gt 1\text{.}\)

11.3.5.

Scratchwork.
Since we are trying to show that there are no integers satisfying , it may be useful to use proof by contradiction.
So, assume \(\exists x,y\in\mathbb Z\) such that \(5y^2-4x^2=7\text{.}\) This may not look too helpful since we have \(2\) unknowns and one equation, and it is not obvious as to how \(x\) and \(y\) would interact with each other to produce \(7\text{.}\) One thing we can do in this question is to try to find a way to reduce the number of the variables. This may look a little unconventional, but if we look at the equation modulo \(4\text{,}\) we see that it becomes
\begin{align*} 5y^2-4x^2 \amp \equiv 7 \pmod 4\\ 1y^2-0x^2 \amp \equiv 3\pmod 4\\ y^2 \amp \equiv 3 \pmod 4 \end{align*}
which eliminates \(x\) completely!
Then, if we can show that there is no integer \(y\) satisfying \(y^2\equiv 3 \pmod 4\text{,}\) we would get our contradiction. Moreover, we see that if \(y\) is even, then \(y^2\equiv 0 \pmod 4\) and if \(y\) is odd, then \(y^2\equiv 1 \pmod 4\text{.}\) Therefore, we have our contradiction.
Notice that we could also consider the four classes of \(y \pmod 4\text{,}\) but it will give the same result, just with more work.
We can also eliminate \(y\) instead of \(x\) by considering the equation modulo \(5\text{.}\) This gives
\begin{gather*} 5y^2 - 4 x^2 \equiv 7 \pmod 5\\ 0y^2 + 1 x^2 \equiv 2 \pmod 5 \end{gather*}
So we need to find \(x^2 \equiv 2 \pmod 5\text{.}\) Writing the square modulo \(5\) gives the following table:
\begin{align*} \begin{array}{c|c|c|c|c|c} x \amp 0 \amp 1 \amp 2 \amp 3 \amp 4\\ \hline x^2 \amp 0 \amp 1 \amp 4 \amp 4 \amp 1 \end{array} \end{align*}
So there is no \(x\) so that \(x^2 \equiv 2 \pmod 5\text{.}\)

11.3.6.

Scratchwork.
  1. We’ll prove the statement by contradiction, and assume that there is some smallest positive rational number, say \(r\text{.}\) We’ll try to find a positive rational number that’s strictly less than \(r\text{.}\) Let’s consider \(r/2\text{,}\) which we know satisfies the inequality \(0\lt r/2\lt r\text{.}\) This inequality isn’t enough though; we also need to show that \(r/2\) is a rational number, and then we’ll have reached a contradiction.
  2. The proof is very similar to part (a). We’ll assume that we have a smallest positive irrational number, say \(r\text{.}\) We’ll try to show that \(r/2\) is irrational. We can prove this statement by using contradiction as well: assume that \(r/2\) is rational, and show this contradicts the assumption that \(r\) is irrational.

11.3.8.

Scratchwork.
This is a good proof-by-contradiction question. We’ll assume that \(\sqrt[3]{25}\) is rational, write it as a simple fraction, and then show that we get a contradiction.

11.3.10.

Scratchwork.
Let’s assume the conclusion is false, that is, that \(rx\) is rational. Then we can write \(rx=p/q\) and \(r=m/n\) for integers \(p,q,m,n\text{.}\) Using this, we can show that \(x\) is rational, which is a contradiction. However, when we write the proof up, we need to be careful to always show that our denominators are non-zero

11.3.11.

Scratchwork.
  1. We have to be careful when dealing with the not equal signs, because the property of being not equal is not transitive. It can be very easy to reach faulty conclusions. For example it is certainly not the case that \(2 \neq -2\) implies that \((2)^2 \neq (-2)^2\text{.}\)
    In the same way, the statement that \(x\neq m/n\) for all \(m,n\in\mathbb{Z}\) and \(y\neq p/q\) for all \(p,q\in\mathbb{Z}\) does not imply that \(xy\neq (mp)/(nq)\) for all \(m,p,n,q\in\mathbb{Z}\text{.}\) This statement isn’t true because we can factor an integer into irrational components; for example, we can write \(2=\sqrt{2}\cdot\sqrt{2}\text{.}\)

11.3.13.

Scratchwork.
  1. We know that the product of two odd numbers is odd, from Exercise 3.5.2. From this, we know that if \(5^k\) is odd, then \(5^{k+1}\) will be too. So we can prove the statement for all \(k\) by using induction.
  2. We’ll prove the statement by contradiction, by assuming that the statement is false, that is, \(\log_2(5)\) is rational. Since \(\log_2(5)=m/n\) we know that \(5 = 2^{m/n}\) and so \(5^n=2^m\text{.}\) But this means that \(5^n\) is even which gives a contradiction. We need to be a little careful with the signs of \(m,n\text{.}\)
  3. Let \(n\in\mathbb{N}\text{.}\) As given in the question, there is some \(a\in \mathbb{Z}\text{,}\) \(a\geq 0\) and \(b\in \mathbb{Z}\) that is odd, so that \(n=2^ab\text{.}\) We know that
    \begin{equation*} \log_2(n)=\log_2(2^ab)=\log_2(2^a)+\log_2(b)=a+\log_2(b). \end{equation*}
    Since \(a\) is an integer, and so rational, the rationality or irrationality of \(\log_2(n)\) will depend on that of \(\log_2(b)\text{.}\) Indeed, we know that
    This implies, respectively, that
    • \(\log_2(n)\) is rational if \(\log_2(b)\) is rational;
    • \(\log_2(n)\) is irrational if \(\log_2(b)\) is irrational.
    We can try to adapt the argument from part (b) to show that \(\log_2(b)\) is irrational, since (b) is a special case of this, with \(b=5\text{.}\) However, in part (b), we used the fact that \(\log_2(5)\gt0\text{,}\) which is only true for \(\log_2(b)\) if \(b\neq 1\text{.}\) We can try to prove that \(\log_2(b)\) is irrational when \(b\neq 1\) and \(b\) is odd.
    The last case to analyze is if \(b=1\text{.}\) Then \(\log_2(b)=0\text{,}\) and hence it is rational. To summarize, we have shown that
    • if \(b=1\text{,}\) then \(\log_2(b)\) is rational, and so \(\log_2(n)\) is rational;
    • if \(b\neq1\text{,}\) then \(\log_2(b)\) is irrational, and so \(\log_2(n)\) is irrational.
    These two statements imply that for \(n\in\mathbb{N}\text{,}\) \(\log_2(n)\) is irrational if and only if \(n\) is not a power of \(2\text{.}\)

11.3.15.

Scratchwork.
We’ll prove this by contradiction by first assuming that there are some \(a,n\in\mathbb N\) satisfying the equation. Writing \(a^2=7^n-35\text{,}\) we can show that \(a^2\) is divisible by \(7\text{.}\) By Euclid’s lemma, this implies that \(7\mid a\text{.}\) Then \(7^2\mid a^2\text{.}\) Notice that \(7\mid 35\text{,}\) but \(7^2\nmid 35\text{.}\) If we also had \(7^2\mid 7^n\text{,}\) then that would imply \(7^2\) divides \(35=7^n-a^2\text{,}\) which is a contradiction. So we have a contradiction in the case \(n\geq 2\text{.}\) Now we just need to show that the case \(n=1\) also leads to a contradiction. In this case, we have the equation \(a^2+35=7\text{.}\) This is impossible, since \(a^2+35\geq 35\text{.}\)

11.3.16.

Scratchwork.
Let’s explore the inequality a little bit before we leap into proving it. Notice that because \(x \in (0,1)\) we know that the denominator is positive. Therefore we can rearrange the inequality to give
\begin{equation*} 1 \geq 4x-4x^2, \end{equation*}
which we can rearrange further to give us
\begin{equation*} (2x-1)^2=4x^2-4x-1 \geq 0. \end{equation*}
This looks good because it tells us that a square is non-negative and we know that squares are non-negative. That smells like the direct proof, what about contradiction?
We need to have \(x\) so that the opposite inequality holds. That is
\begin{equation*} \frac{1}{2x(1-x)} \lt 2, \end{equation*}
but this gives us (by similar reasoning) that
\begin{equation*} 1 \lt 4x-4x^2 \end{equation*}
and so
\begin{equation*} (2x-1)^2= 4x^2-4x+1 \lt 0. \end{equation*}
This smells like a contradiction since it tells us a square is negative.

11.3.17.

Scratchwork.
  1. Let’s try to rearrange the inequality in the statement to achieve something more obvious.
    \begin{align*} \frac{x}{y}+\frac{y}{x} \amp \gt 2. \end{align*}
    We give the left hand side a common denominator
    \begin{align*} \frac{x^2+y^2}{xy} \amp \gt 2. \end{align*}
    Now we multiply both sides by \(xy\text{.}\) This won’t change the sign of our inequality because \(x, y \gt 0\text{,}\) so \(xy \gt 0\)
    \begin{align*} x^2+y^2 \amp \gt 2xy. \end{align*}
    We now subtract \(2xy\) from both sides of the equation and see if the resulting quadratic equation has any nice properties
    \begin{align*} x^2-2xy+y^2 \amp \gt 0. \end{align*}
    We factor the left hand side
    \begin{align*} (x-y)^2 \amp \gt 0. \end{align*}
    Since we assumed that \(x \neq y\text{,}\) we know that \(x-y \neq 0\text{,}\) and the square of any non-zero real must be positive. Therefore we’ve reached an obvious fact! To write the direct proof, we start with the obvious fact and work backwards through the steps above until we reach the desired result.
  2. We check what happens when \(x=y\text{.}\) When \(x=y\text{,}\) \(\frac{x}{y} = \frac{y}{x} = 1\text{.}\) Thus,
    \begin{equation*} \frac{x}{y} + \frac{y}{x} = 2. \end{equation*}
    We’ll alter the statement to read:
    For all \(x, y\in \mathbb{R}\) with \(x, y \gt 0\)
    \begin{equation*} \frac{x}{y}+\frac{y}{x} \geq 2. \end{equation*}

11.3.18.

Scratchwork.
Consider the expression needed for a proof by contradiction. Let \(a,b \gt 0\text{.}\) What would it mean for us to have
\begin{equation*} \frac{2}{a} + \frac{2}{b} = \frac{4}{a+b}? \end{equation*}
Multiply both sides by \(ab(a+b)\) to clear the denominators
\begin{equation*} 2a^2+2b^2 +4ab=4ab, \end{equation*}
and thus
\begin{equation*} a^2+b^2 = 0. \end{equation*}
But since \(a,b \gt 0\text{,}\) \(a^2+b^2 \gt 0\) and we’ve obtained a contradiction!
We can leverage this reasoning into a direct proof too. We leave that to the reader.

12 Cardinality
12.7 Exercises

12.7.3.

Scratchwork.
We know that we can use this function to show that \(\mathbb N\times\mathbb N\) is denumerable, if it is a bijective function. That is, if the function is bijective, then we can conclude that \(|\mathbb N|=|\mathbb N\times\mathbb N|\text{,}\) which implies that \(\mathbb N\times\mathbb N\) is denumerable. But, if the function is not bijective, this wouldn’t give us any more information, since just because \(f\) is not bijective wouldn’t imply \(|\mathbb N|\neq|\mathbb N\times\mathbb N|\text{.}\)
We actually saw that if we have two denumerable sets \(A, B\text{,}\) then \(A\times B\) is also denumerable. We proved it using a diagonalisation argument (see Result 12.2.6). Applying this result to \(A=B=\mathbb N\text{,}\) we can immediately see that \(\mathbb N\times \mathbb N\) is, indeed, denumerable.
So, assuming that we can show \(f\) is bijective, this question suggests that we could also show that \(\mathbb N\times \mathbb N\) is denumerable directly by finding a bijective function from \(\mathbb N\times\mathbb N\) to \(\mathbb N\text{.}\)

12.7.6.

Scratchwork.
  1. Since we’re trying to figure out whether or not \(\mathbb{I}\) is denumerable, let’s assume that it is, and see what kind of conclusion we may reach. That is, let’s see if under this assumption we reach a contradiction.
  2. We know that
    \begin{equation*} [0,1]\cap \mathbb{Q}\subset \mathbb{Q}, \end{equation*}
    and that \(\mathbb{Q}\) is denumerable. Since any subset of a denumerable set is countable, and the subset is infinite, it must be denumerable.
  3. The set
    \begin{equation*} \set{\pi+q:q\in\mathbb{Q}} \end{equation*}
    shifts the set of rational numbers by \(\pi\text{.}\) Intuitively, we think that shifting shouldn’t change the size of a set, but of course, we need to prove this by showing that the set is denumerable, like \(\mathbb{Q}\text{.}\) Indeed, we can define a bijection between \(\set{\pi+q:q\in\mathbb{Q}}\) and \(\mathbb{Q}\text{.}\) Since \(\mathbb{Q}\) is denumerable, the other is too.
  4. Part (c) is a special case of this, with \(a=\pi\text{.}\) We need to check that the argument given in part (c) works when \(a\) is an arbitrary real number.
  5. The set
    \begin{equation*} \set{\pi q:q\in\mathbb{Q}} \end{equation*}
    scales the set of rational numbers by \(\pi\text{.}\) In comparison to part (c), it may not be as intuitive that this set has the same size as the rational numbers. However, we can define the bijection \(g(q)=\pi q\) from the rational numbers to \(\set{\pi q:q\in\mathbb{Q}}\text{.}\)
  6. Part (e) is a special case of this, with \(a=\pi\text{.}\) We need to check that the argument given in part (e) works when \(a\) is an arbitrary real number.

12.7.8.

Scratchwork.
The hints suggests that we think about a bijection from \((0,\infty)\) to \(\mathbb R\text{.}\) We see that the easiest of such function is \(f(x)=\log(x)\text{.}\) But, we also realize that the set given to us is not \((0,\infty)\text{,}\) but \((-\infty, -\sqrt{29})\text{.}\) This may look like a big problem, but it is a problem we can easily overcome. First observe that we can take \(g(x)=\log(-x)\text{,}\) instead of \(f\text{,}\) it becomes a bijective function (at least intuitively, we need to show that such changes preserve bijectivity) from \((-\infty,0)\) to \(\mathbb R\text{.}\) Then, we can finalize our construction by shifting \(g\) by \(\sqrt{29}\) to the left, i.e. by defining \(h(x)=\log(-x-\sqrt{29})\text{.}\) This function, \(h\text{,}\) should be a bijection from \((-\infty, -\sqrt{29})\) to \(\mathbb R\text{.}\)
Now, of course, we need to show that \(h\) is indeed a bijection. We can do this either by showing that it is injective and surjective, or by showing that it has both a left and a right inverse. Since we know how to find the inverse of the \(\log\) function, the second way may be easier.

12.7.9.

Scratchwork.
Question 13 in Section 10 suggests that we should be able to construct a bijective function \(g: S\to\pow{N}\) (as done in that question) implying that \(S\) is uncountable.

12.7.10.

Scratchwork.
  • First solution: Let’s try to find a bijection, say \(h\text{,}\) from \((0,1)\) to \(\mathbb{R}\text{.}\) We’ll need \(h\) to blow up to infinity at some point, and negative infinity at another point. Let’s try to create \(h\) so that
    \begin{equation*} \lim_{x\to 0^+}h(x)=+\infty, \quad \lim_{x\to 1^-}h(x)=-\infty. \end{equation*}
    We know that the function \(1/x\) satisfies the first condition, and that the function \(1/(x-1)\) satisfies the second condition.
    We can define \(h\) to be a piecewise function involving rational functions like these. Suppose we tried to define
    \begin{equation*} h(x)=\begin{cases} \frac{1}{x} \amp \text{if } 0\lt x \leq \frac{1}{2} \\ \frac{1}{x-1} \amp \text{if } \frac{1}{2}\lt x\lt 1 \end{cases} \end{equation*}
    This function doesn’t quite work, since it jumps from \(2\) to \(-2\) at \(x=1/2\text{,}\) and is thus not surjective. Let’s fix this by defining
    \begin{equation*} h(x)=\begin{cases} \frac{1}{x}-2 \amp \text{if } 0\lt x \leq \frac{1}{2} \\ \frac{1}{x-1}+2 \amp \text{if } \frac{1}{2}\lt x\lt 1 \end{cases} \end{equation*}
    We now need to show that this is a bijection.
  • Second solution: Instead of working with the piecewise function, we could also try the function \(y:(0,1)\rightarrow\mathbb{R}\) defined by
    \begin{equation*} y(x) = \frac{1}{x}-\frac{1}{1-x}=\frac{1-2x}{x(1-x)} \end{equation*}
    which also blows up to positive infinity as \(x\) approaches \(0\) from the right, and blows up to negative infinity as \(x\) approaches \(1\) from the left. We need to prove that it is bijective.
  • Third solution: We can show that both sets are equinumerous by showing they are both equinumerous to the interval \((1,\infty)\) via simpler bijections.

12.7.12.

Scratchwork.
In this question, we intuitively know that the result must be true. This is because we realize that if all the sets are finite, the result is more or less trivial. When the sets are finite we know that the we can equate the cardinality of a set with its “size”, i.e. the number of elements it has. This means that if \(|A|=|B|\) and \(|C|=|D|\text{,}\) we can conclude that \(A\) and \(B\) have the same number of elements and \(C\) and \(D\) have the same number of elements. Thus, since \(A\cap C=\emptyset\) and \(B\cap D=\emptyset\text{,}\) we get
\begin{equation*} |A\cup C|=|A|+|C|=|B|+|D|=|B\cup D|, \end{equation*}
and the result follows.
Unfortunately, this argument fails when at least one of the sets is infinite, since we cannot talk about the “sizes” of infinite sets. So what we need to, instead, is to use the actual definition of two sets having equal cardinalities. That is, we know that \(|A|=|B|\) and \(|C|=|D|\) means that there exist two bijective functions \(f: A\to B\) and \(g: C\to D\text{.}\)
Now, we need to be able to find a bijective function
\begin{equation*} h:A\cup C \to B\cup D. \end{equation*}
For this, we realize that if \(x\in A\cup C\text{,}\) then \(x\in A\) or \(x\in C\text{.}\) This means that for each case, we can use either \(f\) or \(g\) to send \(x\in A\cup C\) to \(B\cup D\text{,}\) which will define our function \(h\text{.}\) Then all we will have to do is to show that this new function is indeed bijective.

12.7.13.

Scratchwork.
We see that if we only want to show that these two sets are equinumerous, our best choice would be to use The Cantor-Schröder-Bernstein theorem. We can do that by finding two injective functions:
\begin{equation*} g_1:(0,\infty)\to (0,\infty)-\set{1}, \end{equation*}
and
\begin{equation*} g_2:(0,\infty)-\set{1}\to (0,\infty). \end{equation*}
In this case, we see that we can take \(g_1\) and \(g_2\) defined as \(g_1(x)=x+1\text{,}\) and \(g_2(x)=x\text{.}\) We see that these functions are injective which implies that \(|(0,\infty)|= |(0,\infty)-\set{1}|\text{.}\)
But, in this question, we want to find an explicit bijection from \((0,\infty)\) to \((0,\infty)-\set{1}\text{.}\) The problem with such bijections is that it generally feels almost trivial since the two sets are almost identical. This gives the impression that we should be just take the identity function, or a tiny variation of it to get the bijection.
However, this idea fails when we realize that our bijection should send 1 to an element in \((0,\infty)-\set{1} \text{,}\) call it \(y_1\text{.}\) But then \(y_1\in (0,\infty)\) and thus should go to another element in the codomain (different than itself), call it \(y_2\text{.}\) Then, this \(y_2\) should go to another element, and that should go to another, and so on. This means that we cannot take a simple variation of the identity function here, we may need to make a more drastic change to account for this “snowball” effect.
The simplest way to handle this situation is by defining a function that sends \(1\) to \(2\text{,}\) \(2\) to \(3\text{,}\) \(3\) to \(4\text{,}\) and so on, very similar to “Hilbert’s Hotel Problem”. As for elements that are not natural numbers, we can just send them to themselves.

12.7.14.

Scratchwork.
  1. Cantor-Schröder-Bernstein 12.5.1 tells us that it is enough to find an injection from \((0,1) \to (0,1)^2\) and then another injection back. The first injection is easy, we can take, say, the function \(f:(0,1) \to (0,1)^2\) defined by \(f(x) = (x,1/2)\text{.}\)
    The injection back is harder; we need to make a pair of reals \((a,b)\) to a single real \(c\text{.}\) A nice “trick” uses the decimal expansions of \(a,b\text{.}\) In particular, we can write \(a = 0.a_1a_2a_3\dots, b=0.b_1b_2b_3\dots\text{.}\) We can then interleave these expansions to create a new number \(c = 0.a_1b_1a_2b_2a_3b_3\dots\text{.}\) Some care is required to ensure that this mapping is injective.
  2. We already know that \(|(0,1)| = |\mathbb{R}|\) and so we know that there is a bijection between these sets. We can then leverage that bijection to construct a bijection between \((0,1)\times(0,1)\) and \(\mathbb{R}^2\text{.}\) Then part (a) tells us that \(|(0,1)\times(0,1)| = |(0,1)|\text{.}\) And then again we can use \(|(0,1)|=|R|\text{.}\) Putting all these together gives us the required result.

12.7.15.

Scratchwork.
Since \(A\) and \(B\) are equinumerous, there is a bijection \(f:A\rightarrow B\text{.}\) We can use \(f\) to create a bijection from the power set of \(A\) to the power set of \(B\text{.}\) Let \(F:\mathcal{P}(A)\rightarrow \mathcal{P}(B)\) be defined by
\begin{equation*} F(C)=\{f(a):a\in C\}. \end{equation*}
Notice that this function \(F\) takes as input a subset of \(A\text{,}\) and maps it to a subset of \(B\text{.}\) We need to show that \(F\) is surjective and injective.

12.7.17.

Scratchwork.
We saw in Exercise 12.7.12 that if \(A_1, A_2, B_1, B_2\) are any nonempty sets satisfying \(A_1\cap A_2=\emptyset\) and \(B_1\cap B_2=\emptyset\text{,}\) and that \(|A_1|=|B_1|\) and \(|A_2|=|B_2|\text{,}\) then we have \(|A_1\cup A_2|=|B_1\cup B_2|\text{.}\) The idea in the proof relied on the fact that since \(|A_1|=|B_1|\) and \(|A_2|=|B_2|\text{,}\) there are bijections \(f:A_1\to B_1\) and \(g: A_2\to B_2\text{.}\) Using these functions, we could construct a piecewise function \(h:A_1\cup A_2 \to B_1\cup B_2\) that turned out to be a bijection as well.
We can see this result as a specific example to this question, where our partitions for the sets \(A=A_1\cup A_2\) and \(B=B_1\cup B_2\) are given as
\begin{equation*} P=\set{A_1, A_2}, \end{equation*}
and
\begin{equation*} Q=\set{B_1, B_2}. \end{equation*}
Then, we see that \(F:P\to Q\text{,}\) defined as \(F(A_1)=B_1\text{,}\) and \(F(A_2)=B_2\) is a bijection from \(P\) to \(Q\text{,}\) and moreover \(|A_1|=|B_1|\) and \(|A_2|=|B_2|\text{.}\)
This suggests that if we have partitions with more sets in them, we should be able to construct a bijection in a very similar piecewise manner. Also, since sets in a partition are mutually disjoint, and that their union gives the entire set, such a construction would still work.

12.7.18.

Scratchwork.
  1. By definition, we know that there is an injection from \(A\) to \(B\text{,}\) and from \(B\) to \(C\text{.}\) We need to use these to create an injection from \(A\) to \(C\text{.}\)
  2. We need to prove that the statement given in the question implies the Cantor-Schröeder-Bernstein Theorem, and conversely, that the Cantor-Schröeder-Bernstein Theorem implies the given statement.
    To prove the first implication, we assume that the statement given in the question is true; that is, if \(|A|\leq |B|\leq |C|\) and \(|A|=|C|\text{,}\) then \(|A|=|B|=|C|\text{.}\) When we take \(A=C\text{,}\) we have the statement of the Cantor-Schroeder-Bernstein Theorem.
    To prove the converse, we need to use the Cantor-Schröeder-Bernstein Theorem to establish the given statement:
    If \(|A|\leq |B|\leq |C|\) and \(|A|=|C|\text{,}\) then \(|A|=|B|=|C|\text{.}\)
    We’ll try to prove that \(|B|=|C|\text{.}\) By Theorem 12.5.1, we can do this by showing that \(|B|\leq |C|\) and \(|C|\leq |B|\text{.}\) The first is given by the statement of the problem. The second we will have to deduce from the assumption that \(|A|=|C|\text{.}\)

12.7.20.

Scratchwork.
  1. Since each \(A_n\) is denumerable, we can list the elements of each set, like
    \begin{equation*} A_n=\set{a_{n1},a_{n2},a_{n3},\dots}. \end{equation*}
    We can then construct an infinite table, where the \(n^{th}\) row of the table lists the elements of \(A_n\text{.}\) Then, just as in the proof of Result 12.2.6, we can construction a bijection from the natural numbers to the elements of the table by sweeping over successive diagonals.
  2. This part differs to part (a), as the sets \(A_n\) may be finite or denumerable. In particular, it’s possible that some of the \(A_n\) are the empty set.
    Initially, suppose that only finitely many of the sets \(A_n\) are non-empty. We’ll return to the other possibility shortly. By relabeling if necessary, we may assume that the first \(N\) sets, \(A_1,\dots,A_N\) are non-empty, and \(A_n=\emptyset\) for \(n\gt N\text{.}\) Then
    \begin{equation*} \bigcup_{n=1}^NA_n = \bigcup_{n=1}^\infty A_n. \end{equation*}
    We are now only taking the union over a finite number of sets, and we need to show that it is countable. Result 12.2.9 tells us that the union of two countable sets is again countable. To get that the union of \(N\) countable sets is also countable, we’ll need to apply induction.
  3. We would like to define sets \(B_n\) such that
    \begin{equation*} \bigcup_{n=1}^\infty A_n=\bigcup_{n=1}^\infty B_n \end{equation*}
    and \(B_m\cap B_n=\emptyset\) for \(m\neq n\text{.}\) That way we may apply part (b) to get the desired result.
    Let \(B_1=A_1\text{.}\) Then, let \(B_2\) consist of all elements of \(A_2\) that are not also elements of \(A_1\text{.}\) That way \(A_1\cup A_2= B_1\cup B_2\text{,}\) and \(B_1\cap B_2=\emptyset\text{,}\) by construction. Next, we would define \(B_3\) to be all elements of \(A_3\) that are not elements of \(A_1\) or \(A_2\text{.}\) Another way of writing this is
    \begin{equation*} B_3 = A_3\setminus(A_1\cup A_2). \end{equation*}
    We proceed in this manner to define \(B_n\) for all \(n\in\mathbb{N}\text{.}\) We need to show that the union of all the sets \(B_n\) is equal to the union of all the sets \(A_n\text{,}\) and that the sets \(B_n\) satisfy the conditions of part (b).