0% found this document useful (0 votes)
31 views

Typed Notes - Real Analysis

This document provides an overview of a real analysis course. It will cover set theory and the real number system through the following topics: [1] Peano's axioms for defining the natural numbers; [2] Zermelo-Fraenkel set theory axioms for defining sets; [3] Proving that the natural numbers form a Peano set using these axioms. The course aims to lay a rigorous mathematical foundation for calculus rules by proving properties that are commonly taken for granted.
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
31 views

Typed Notes - Real Analysis

This document provides an overview of a real analysis course. It will cover set theory and the real number system through the following topics: [1] Peano's axioms for defining the natural numbers; [2] Zermelo-Fraenkel set theory axioms for defining sets; [3] Proving that the natural numbers form a Peano set using these axioms. The course aims to lay a rigorous mathematical foundation for calculus rules by proving properties that are commonly taken for granted.
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 48

UM 101 (FALL 2022)

PART A: Real Analysis

What is this course about? This is “calculus for thinking practitioners". We will lay the founda-
tion for “rules" that you are already very good at. So, we may work very hard to prove something
that you already know, but it is the process which is the focus of the course. You will enjoy it if you
are curious about “why" certain rules work, or “when do they fail", which is the kind of curiosity
that makes a mathematician. For others, think of this as a training in rigor that will prepare you
for the rest of this program.

1. S ET THEORY AND THE REAL NUMBER SYSTEM

1.1. The ZFC & Peano axioms. To do analysis on real and complex numbers, we must talk about
what those really are. And rather than take them completely for granted, we must understand
where they come from. For that, we must understand rational numbers, which come from in-
tegers, which themselves come from natural numbers. Is it possible to synthesize a basic set of
rules that eventually yield all the properties of natural numbers that we have been happily using?
Peano (1858-1932) tried to do this:

Definition 1.1 (Peano’s axioms). A set A, is called a Peano set if it satisfies the following axioms or
“rules".
(P1) A contains a distinguished element, which is called 0. In particular A is non-empty.
(P2) There exists a function S from A into itself.
(P3) S(a) ̸= 0 for any a ∈ A.
(P4) S is injective, i.e., if S(a) = S(b) for any a, b ∈ A, then a = b.
(P5) (Principle of Mathematical Induction) if B is a subset of A such that 0 ∈ B , and S(a) ∈ B
whenever a ∈ B , then B = A.
The function S is called a successor function on A.

Class discussion. Explain why the following choices of A and S do not make A a Peano set. You
must explicitly mention which of the above axioms are failing.
a. A = {0, 1}, S(0) = 1 and S(1) = 1
b. A = {0, 0.5, 1, 1.5, 2, 2.5, ...} and S(x) = x + 1.
1
2

It is “clear" to us that the “usual" natural numbers {0, 1, 2, ...} endowed with S(n) = n +1 is a Peano
set. One could take this for granted and proceed ahead (as we have done for many years in
school). If you take only the existence of a Peano set for granted, you may argue as follows: why
don’t we define

1 := S(0);
2 := S(S(0))

and so on. Then using the Principle of Mathematical Induction, we show that the original Peano
set contains nothing extra, and we are almost there (barring the issue of defining addition). How-
ever, this approach forces use to understand what “so on" (definition by recursion) means, and
why the above procedure yields a subset and so on. In fact, we have to ascribe meaning to all
of the italicized words above. In short, we must describe the fundamental rules governing the
formation of sets and functions!

Definition 1.2. A set is a well-defined collection of (mathematical) objects that are called ele-
ments of the set. If A is a set and x is an object, then we use the notation

x∈A

to denote the fact that x is an element of A. Otherwise, we write

x ∉ A.

Given two sets A, B , we say that A is a subset of B , or A ⊆ B , if every element of A is an element


of B . Otherwise, we write that A ̸⊆ B . This is different from A ⊊ B , i.e., A is a proper subset which
means that every element of A is an element in B , but there is an element in B which is not in A.

Example 1. Going back to our known understanding of sets, let A = {1, 2, {3}}. Then, 2 ∈ A, {2} ⊆ A,
but {2} ∉ A. On the other hand, {3} ∈ A but 3 ∉ A, and therefore, {3} ̸⊆ A.

END OF LECTURE 1
The ZFC Axioms - “Naive" approach. Note that not every collection of mathematical objects is
considered a set (read about Russell’s Paradox)! The following rules determine what is and isn’t
allowed to be called a set. (There are some redundancies in this list of axioms.)
A. Basic. Every mathematical object is a set. (This means that if you know that something is
an element of a set, then that element itself is considered to be a set.)
B. Extension. Two sets, A and B , are said to be equal, i.e., A = B , if they have the same
elements. Practically, to show A = B , one often shows that A ⊆ B and B ⊆ A.
3

Example 2. Show that the following two sets are equal

A = {0, 1},
B = the set of possible remainders obtained when
dividing a perfect square by 4.

Let a ∈ A, we will show that there is a natural number c such that c 2 ≡ a (mod 4). For
a = 0, take c = 2. For a = 1, take c = 1. Thus, A ⊆ B . On the other hand, let b ∈ B . Then,
b = c 2 − 4q for some natural numbers c and q, and 0 ≤ b ≤ 3. If c is even, then 4 divides
b, which can only happen if b = 0. If c is odd, i.e., c = 2k + 1, then b = (2k + 1)2 − 4q =
4(k 2 +k − q)+1. The only number in {0, 1, 2, 3} of the form 4· +1 is 1. Thus, we have shown
that B ⊆ A.

C. Existence. There exists a set with no elements, called the empty set, denoted by ;.
D. Specification. Let A be a set and P (a) be a “property" that applies to every a ∈ A, i.e., P (a)
is either true or false, then

B = {b ∈ A : P (b) is true}

exists as a set and is a subset of A.

Example 3. B = {a ∈ N : 4|a} is the set of natural numbers divisble by 4.

E. Pairing. Given two sets A, B , there exists a set containing exactly A and B as its elements,
which is denoted by {A, B }. Note that this also gives the existence of {A}.
F. Union. For a set F of sets, there exists a set, called the union of the sets in F , whose
elements are precisely the elements of the elements of F . I.e.,

A ⇐⇒ x is an element of at least one A ∈ F .


[
x∈
A∈F

Consequence. 1. For a non-empty set F of sets, there exists a set, called the intersection
of the sets in F such that

A ⇐⇒ x is an element of everyA ∈ F .
\
x∈
A∈F

G. Power Set. Given a set A, there exists a set P (A) whose elements are precisely the subsets
of A. This set is called the power set of A.

Remark. The axioms so far give the existence of unions, intersections (why?), set differ-
ences, and (finite) Cartesian products (this is more complicated) as sets. In particular,
given two sets A, B , the set A ×B is the set of all ordered pairs (a, b), where a ∈ A and b ∈ B .
4

END OF LECTURE 2
Definition 1.3. A relation from A to B is a subset R of A × B . We say that aRb if and only
if (a, b) ∈ R. The domain of R is

dom(R) = {x ∈ A : (x, y) ∈ R for some y ∈ B }.

The range of R is

ran(R) = {y ∈ B : (x, y) ∈ R for some x ∈ A}.

A function from A to B is a relation f ⊂ A × B such that dom( f ) = A and for each x ∈ A,


there is at most one y ∈ B such that x f y. This y is denoted by f (x).

H. Replacement. Skip!
I. Regularity. Skip!
J. Choice. Skip!

Definition 1.4. Given a set A, the successor of A is the set A + = A ∪ {A}. A set which
contains the empty set as an element and the successors of each of its elements is called
an inductive set.

Question. Why does the A + exists?

K. Infinity. There exists an inductive set.

Lemma 1.5. Let C be a non-empty set of inductive sets. Then, B =


T
A∈C A is an inductive set.

Proof. Homework! □

Theorem 1.6. There exists a unique minimal inductive set. That is, there is an inductive set ω such
that ω ⊆ S for every inductive set S, and if ω′ is an inductive set with the same property, then ω = ω′ .

Proof. Let A be an inductive set, whose existence is guaranteed by the Axiom of Infinity. By the
Axiom of Power Set, P (A) is a set. Then, by the Axiom of Specification,

F = {W ∈ P (A) : W is an inductive set}

is a set. Since F is non-empty (why?), we may consider

ω=
\
W.
W ∈F
Step 1. ω is inductive. This is a direct consequence of Lemma 1.5.
5

Step 2. ω is minimal. Let S be an inductive set. Then, by Lemma 1.5, S ∩ A is an inductive set. By
the definition of F , S ∩ A ∈ F . By the definition of intersections, ω ⊆ S ∩ A ⊆ S. Thus, ω ⊆ S (do you
know what this property of ⊆ is called?).
Step 3. ω is unique. By minimality, ω ⊂ ω′ and ω′ ⊂ ω. By the Axiom of Extension, ω = ω′ . □

Theorem 1.7. Let ω be the minimal inductive set. Set 0 := ; and S(A) = A + . Then, ω is a Peano
set.

Theorem 1.8 (The recursion principle). Let X be a non-empty set, f : X → X be a function, and
a ∈ X . Then, there exists a unqiue function F : ω → X such that
(a) F (;) = a
(b) F (A + ) = f (F (A)) for each A ∈ ω.

Remember that the above theorem is claiming the existence of a particular subset of ω× X . When
we define sequences recursively, we are essentially applying the above theorem, as we will see
shortly.

1.2. The Natural numbers. Within the ZFC regime, ω is defined as the set of natural numbers.
But, how do we recover the familiar set of natural numbers from the above discussion? We iden-
tify N and ω as follows

0 := ;
1 := 0+ = {;} = {0}
2 := 1+ = {;, {;}} = {0, 1}
3 := 2+ = {;, {;}, {;, {;}}} = {0, 1, 2}

and so on.
END OF LECTURE 3
Definition 1.9 (Peano Addition). Given m ∈ N, the recursion principle gives the existence of a
unique function summ : N → N such that
(1) summ (0) = m;
(2) summ (n + ) = (summ (n))+ .
Define, for m, n ∈ N,

m + n := summ (n).

Let us prove an “obvious" statement!


6

Proposition 1.10. 2 + 3 = 5.

Proof. We start from the left-hand side.

2 + 3 = sum2 (3) = sum2 (2+ ) = (sum2 (2))+ = (sum2 (1+ ))+


= ((sum2 (1))+ )+ = (((sum2 (0))+ )+ )+ = ((2+ )+ )+ = (3+ )+ = 4+ = 5.

Trying to prove 2 + 8 = 10 would be quite tedious, but not so much if we could say that 2 + 8 =
8 + 2. Let us recover the basic properties of addition that we have taken for granted.

Remark. Note that m + = (summ (0))+ = summ (0+ ) = m + 1 by this definition. So will now always
write m + 1 instead of m + .

Definition 1.11 (Peano Multiplication.). Given m ∈ N, the recursion principle gives the existence
of a unique function prodm : N → N such that
(1) prodm (0) = 0;
(2) prodm (n + ) = prodm (n) + m.
Define, for m, n ∈ N,

m · n := prodm (n).

Theorem 1.12. The following hold.


(1) (Commutativity) m + n = n + m and m · n = n · m for all m, n ∈ N.
(2) (Associativity) m + (n + k) = (m + n) + k and m · (n · k) = (m · n) · k for all m, n, k ∈ N.
(3) (Distributivity) m · (n + k) = (m · n) + (m · k) for all m, n, k ∈ N.
(4) m + n = 0 implies that m = n = 0 for m, n, k ∈ N.
(5) m · n = 0 implies that either m = 0 or n = 0 for m, n ∈ N.
(6) (Cancellation) m + k = n + k implies that m = n, and m · k = n · k with k ̸= 0 implies that
m = n for m, n, k ∈ N.
(7) SKIP or Give as exercise!(Order) Given m, n ∈ N, exactly one of the following three statements
hold.
(a) m = n;
(b) there is a k ∈ N \ {0} such that m + k = n;
(c) there is a k ∈ N \ {0} such that n + k = m.
(8) SKIP or Give as exercise! (Order preservation) Given m, n, k ∈ N such that n > m and k > 0,
n · k > m · k.
7

Proof. We will only prove the additive part of (6). The rest are for self-study. Note that subtraction
is not available to us! We will fix an arbitrary m, n ∈ N and prove the statement by induction on k.
Consider the case k = 0. We are given that summ (0) = sumn (0), which by definition gives that
m = n. Next, assume the statement holds for k. We must prove it for k + 1. For this, suppose
m + (k + 1) = n + (k + 1). I.e., summ (k + 1) = sumn (k + 1). By definition, this gives that

summ (k) + 1 = sumn (k) + 1.

But the successor map is injective. So, summ (k) = sumn (k). By the induction hypothesis, m = n.
By the principle of mathematical induction, the cancellation law holds for all k ∈ N. Since m, n ∈
N were arbitrary, the claim holds. □

SKIP or Give as exercise! A natural number n ∈ N is said to be positive if n ̸= 0. We say that n is


greater than m if (b) holds, and write n > m. If either n > m or n = m, we write n ≥ m. Question
Why does the positivity of n imply that n > 0?
END OF LECTURE 4

1.3. Field axioms and ordered fields. The set of natural numbers are missing many nice opera-
tions such as subtraction and division, for instance, we cannot solve

x + 3 = 2,
3x = 5

within the set of natural numbers. In fact, we would like to work with number systems where we
can even solve x 2 + 2 = 0 and x 2 − 2 = 0! First, we introduce some abstract concepts that capture
some of these ideas.

Definition 1.13. A set F with two operations + : F × F → F and · : F × F → F is said to be a field if


(F, +, ·) satisfies the following “field axioms":
(F1) + and · are commutative on R.
(F2) + and · are associative on R.
(F3) + and · satisfy the distributive law on R.
(F4) There exists two distinct elements, denoted by 0 (the additive identity) and 1 (the multi-
plicative identity), such that for all x ∈ F ,

x +0 = x
x · 1 = x.
8

(F5) For every x ∈ F , there is a y ∈ F such that

x + y = 0.

(F6) For every x ∈ F , x ̸= 0, there is a y ∈ F such that

x y = 1.

EXERCISE: Read Theorems 1.1-1.15 from Apostol.

Remark. You should be tempted to denote the y in (F5) by −x and the y in (F6) by 1/x. How-
ever, this notation would be meaningless unless we established the uniqueness of y in both the
axioms. These are Theorems 1.1 and 1.7 in the book!

Remark. Define −x as the unique y given by Theorem 1.1. It is called the additive inverse of x.
We write a + (−b) as a − b for convenience.
Similarly, define 1/x as the unique multiplicative inverse of a nonzero x. We write a · (1/b) as
a/b for convenience.

Theorem 1.14 (Apostol, Theorem 1.6). 0 · x = x · 0 = 0 for all x ∈ F .

Proof. By commutativity of · (F1), we already have the first equality. Now, by (F4), 1 + 0 = 1 and
x · 1 = x. Thus, by (F3)

x = x · 1 = x · (1 + 0) = (x · 1) + (x · 0) = x + (x · 0).

By the cancellation law (Theorem 1.1),


0 = x · 0.

Definition 1.15. A set A with a relation < is called an ordered set if the following holds:
(O1) For every x, y ∈ A, exactly one of the following three holds: x = y, x < y or y < x.
(O2) Order is transitive. Given x, y, z ∈ F , if x < y and y < z, then x < z.
If x < y, we say that x is strictly less than y. The notation x ≤ y denotes the statement x < y or
x = y, and we say that x is less than or equal to y. Similarly, x > y and x ≥ y are interpreted.

Example 4. N is an ordered set if we define < as follows: m < n if there is a k ∈ N \ {0} such that
m + k = n.

Definition 1.16. An ordered field is a set F that admits two operations +, · and a relation < such
that (F, +, ·) is a field, (F, <) is an ordered set and the following conditions hold.
9

(O3) For x, y, z ∈ F , if y < z, then x + y < x + z.


(O4) For x, y ∈ F , if 0 < x and 0 < y, then 0 < x y.

EXERCISE: Read Theorems 1.20-1.25 from Apostol. Our order axioms are DIFFERENT from Apos-
tol’s. His theorems 1.16-1.19 are our order axioms!!

Theorem 1.17 (Apostol, Theorem 1.21). In an ordered field, 0 < 1. We may use Theorem 1.12 from
Apostol: (−a) · b = −(a · b) and (−a) · (−b) = ab.

Proof. From the field axioms, we know that 0 ̸= 1. Thus, by Axiom (O1), either 0 < 1 or 1 < 0. If the
former holds, we are done.
If the latter holds, by (O3), 1 + (−1) < 0 + (−1). Thus, 0 < −1. By (O4), 0 < (−1) · (−1). Thus,
0 < 1, which is a contradiction.
Thus, 0 < 1 hold. □
END OF LECTURE 5
1.4. Upper bound and least upper bound. Let (F, +, ·, <) be an ordered field (you can think of the
usual real numbers, with the usual +, · and < as the key example).

Definition 1.18. A non-empty subset S ⊆ F is said to be bounded above if there is an element


b ∈ F such that
x ≤ b, ∀x ∈ S.
We call b an upper bound of S. An upper bound of S that is contained in S is called a maximum
of S.

Example 5. Let S = {x ∈ F : 0 ≤ x ≤ 1} and T = {x ∈ F : 0 ≤ x < 1}. Both S and T are bounded above.
In both cases, 1 is an upper bound. In the cases, of S, 1 is the maximum of S.

Remark. Upper bounds clearly need not be unique. However, if a bounded set has a maximum,
then the maximum must be unique (why?).

Definition 1.19. Let S ⊆ F be a bounded above set. An element b ∈ F is said to be a least upper
bound of S or supremum of S if
(i ) b is an upper bound of S,
(i i ) for any a ∈ F such that a < b, a is not an upper bound of S. In other words, given a < b,
there exists an s ∈ S such that a < s.

Theorem 1.20. Let S ⊂ F be a bounded set that admits a least upper bound. Suppose b 1 and b 2 are
least upper bounds of S. Then, b 1 = b 2
10

Proof. Note that there are three possiblities: either b 1 = b 2 , b 1 < b 2 or b 2 < b 1 . There is nothing to
prove in the first case.
Suppose b 2 < b 1 . Then, by item (i i ) of Definition 1.19 applied to b 1 , b 2 is not an upper bound
of S. However, by item (i ) of Definition 1.19 applied to b 2 , b 2 is an upper bound of S. This is a
contradiction.
Suppose b 1 < b 2 . The same argument holds with the roles of b 1 and b 2 exchanged. □

Remarks. 1. We denote the least upper bound of a set by sup S.


2. sup S need not belong to S!
3. The maximum of a set, if it exists, is its least upper bound.

Example 6. Let S = {x ∈ R : 0 ≤ x < 1}. Then, sup S = 1.

Proof. As noted earlier, 1 is an upper bound of S.


Now, let a ∈ F such that a < 1. Then, either a < 0 or 0 ≤ a < 1. In the former case, a is not an
upper bound of S since 0 ∈ S and a < 0.
a+1
In the latter case, let s = 2 = 12 · (a + 1). Since 0 < 1, we have 0 ≤ 0 + a < a + 1 (by O2). Thus,
by (O2) 0 < a + 1. Next, by (O4), 0 < s. Next, a < 1. Thus, a + 1 < 2 by (O2). Similar to Problem 4 in
HW 2, s < 1. Thus, s ∈ S. However, we also have that 2a < a + 1. Thus a < s. This says that a is not
an upper bound of S. We have established both (i ) and (i i ) of Definition 1.19. □

1.5. The set of real numbers. We assume the existence of a set R which admits operations +, ·
and a relation < so that (F1)-(F6) and (O1)-(O4) are satisfied and the set admits the least upper
bound property, i.e.,
(LUB) every bounded set S ⊆ R admits a least upper bound.
Some special subsets of R are:
(1) any x > 0 is called a positive number, and any x < 0 is called a negative number,
(2) N = {0, 1, 2, ...}, the set of natural numbers,
(3) P={1,2,...}, the set of positive natural numbers,
(4) Z = N
n ∪ {−n : n ∈ P},othe set of integers,
p
(5) Q = q
: q ∈ P, p ∈ Z , the set of rational numbers,
(6) Q = R \ Q, the set of irrational numbers.

Theorem 1.21 (Archimedean propery). Let x, y ∈ R such that x > 0. There exists a positive integer
n such that nx > y.

Proof. May Skip! Fix x > 0. Let


S = {nx : n ∈ P}.
11

Clearly, S is non-empty. Suppose the claim was not true. Then nx ≤ y for all n ∈ P and y us ab
upper bound of S. Thus, by the l.u.b. property of R. b = sup S exists. Now, since b − x < b, by
item (i i ) of Definition 1.19, there exists an n ∈ P such that b − x < nx. Thus, b < (n + 1)x ∈ S. This
contradicts the fact that b is an upper bound of S. □
END OF LECTURE 6
12

2. S EQUENCES AND SERIES

Going forward, we will allow all the properties of real numbers that we have used so far, in-
cluding exponentiation, the absolute value function, etc.

2.1. Sequences.

Definition 2.1. A sequence in R is a function f : N → R. It is customary to write a sequence as


{a n }∞
n=0 or {a n }n∈N , where
a n := f (n), n ∈ N,
and is called the n th -term of the sequence.

Given a sequence, we would like to rigorously capture the idea that the terms of the sequence
are approaching a single real number.

Definition 2.2. A sequence {a n }n∈N in R is said to be convergent if there exists an L ∈ R such that
for every ε > 0. there exists an Nε,L ∈ N such that

| f (n) − L| < ε for all n ≥ Nε .

In this case, we call L a limit of {a n }n∈N , and write

a n → L as n → ∞.

A sequence {a n }n∈N in R is said to be divergent if it is not convergent.

Theorem 2.3. Let {a n }n∈N be a convergent sequence in R. Suppose ℓ1 and ℓ2 are limits of {a n }n∈N .
Then, ℓ1 = ℓ2 .

Proof. Let ε > 0. By definition, there exists an N1 ∈ N such that |a n − ℓ1 | < ε/2 for all n ≥ N1 .
Similarly, there exists an N2 ∈ N such that |a n − ℓ2 | < ε/2 for all n ≥ N2 . Let N = max N1 , N2 . Then,

0 ≤ |ℓ1 − ℓ2 | ≤ |ℓ1 − a N | + |a N − ℓ2 | < ε.

But, ε > 0 was arbitrary. HW. If 0 ≤ x < ε for all ε > 0, then x = 0.
Note. This proof can also be done via contradiction. □

Remark. We may now safely write


lim a n = L.
n→∞
This statement means that the limit exists and is equal to L.

Example 7. (a) Let p > 0 be fixed. Let a n = 1/n p , n ∈ P. Then, {a n }n∈P is convergent and 0 is a
limit.
13

Proof. Let ε > 0. By the Archimedean property of R applied to x = ε1/p and y = 1, there is an
N = Nε ∈ P such that
N ε1/p > 1.
Thus, for n ≥ N , ¯ ¯
¯ 1 1 1
¯ n p − 0¯ = n p ≤ N p < ε.
¯
¯ ¯

Since ε > 0 was arbitrary, our claim holds. □

END OF LECTURE 7
Example 8. (b) a n = (−1)n , n ∈ N. Then, {a n }n∈N is divergent.

Proof. Suppose {a n }n∈N is convergent and admits a limit L ∈ R. Let ε = 1. By definition, there
exists an N ∈ N such that
|a n − L| < 1 ∀n ≥ N .
Thus, by the triangle inequality,

|a 2N − a 2N +1 | ≤ |a 2N − L| + |a 2N +1 − L| < 1 + 1 = 2.

But the left-hand side is 2. This is a contradiction.


Definition 2.4. A sequence {a n }n∈N is said to be bounded if there exists an M > 0 such that

|a n | < M

for all n ∈ N.

Definition 2.5. A sequence {a n } is said to be (montonically) increasing if a n ≤ a n+1 for all n. A


sequence {a n } is said to be (montonically) decreasing if a n ≥ a n+1 for all n. A sequence is said to
be monotone if it is either increasing or decreasing.

Theorem 2.6. A monotone sequence is convergent if and only if it is bounded.

Proof. (=>) Case 1. Let {a n } be an increasing and bounded sequence. Then, there exists an M > 0
such that
|a n | < M ∀n.
In other words −M < a n < M for all n. Let S = {a n : n ∈ N}. S is nonempty and bounded above. By
LUB, b = sup S exists in R.
Let ε > 0. By the definition of the supremum and monotonicity, ∃N ∈ N such that

b − ε < a N ≤ an ∀n ≥ N .
14

On the other hand, a n ≤ b < b + ε for all n ∈ N. Thus, for all n ≥ N ,

|a n − b| < ε.

But ε was arbitrary.


Case 2. What if {a n } is decreasing?
(<=) In homework 03, you will show that every convergent sequence is bounded! □

" Divergent sequences may diverge for different reasons!

Definition 2.7. We say that a sequence {a n } diverges to ∞ or limn→∞ a n = +∞ if for every R ∈ R,


there exists an NR ∈ N such that a n > R for every n ≥ NR . One can similarly give meaning to
limn→∞ a n = −∞.

Example 9. We show that limn→∞ n = +∞. Let R ∈ R. If R ≤ 0, then n ≥ R for all n ≥ 1. So we may
choose NR = 1. If R > 0, then by the Archimedean property, there is a positive natural number
N such that N > R. Choose NR = N in this case, we get that n > R for all n ≥ NR . Thus, we have
shown that for any R ∈ R, there is an NR ∈ N such that |a n | > R for all n ≥ NR .

Theorem 2.8 (Limit Laws). Let {a n } and {b n } be convergent sequences in R with limits a and b,
respectively.
(1) For any c ∈ R, {a n + c} converges to a + c and {c a n } converges to c a.
(2) The sequence {a n + b n } converges to a + b.
(3) The sequence {a n b n } converges to ab.
(4) Suppose b ̸= 0 and ∃M ∈ N such that b n ̸= 0 ∀n ≥ M then {1/b n }n≥M converges to 1/b.
(5) Suppose b ̸= 0 and ∃M ∈ N such that b n ̸= 0 ∀n ≥ M then {a n /b n }n≥M converges to a/b.

Proof. We will prove (4). Scrapwork: want:


¯ ¯
¯ b − bn ¯
¯ bb ¯ < ε.
¯ ¯
n

If we could show that ¯ ¯


¯ b − bn ¯
¯ bb ¯ < M |b − b n |
¯ ¯
n
for some M independent of n, we would be done. For this, we need an (eventual) lower bound
on |b n |.
Let ε1 = |b|/2. By convergence of {b n } to b, there exists an N1 ∈ N such that, for all n ≥ N1 ,

|b n − b| < ε1 = |b|/2.
15

By the reverse triangle inequality,


¯ ¯
¯|b n | − |b|¯ < |b n − b| < |b|/2.
¯ ¯

Thus,

(2.1) |b n | > |b| − |b|/2 = |b|/2 ∀n ≥ N1 .

This gives that for all n ≥ N1


¯ ¯ ¯ ¯
¯ 1 1 ¯ ¯ bn − b ¯
¯ − ¯=¯ ¯ < 2 |b n − b|.
¯b
n b ¯ ¯ b n b ¯ |b|2
Let ε > 0.
2
Set ε2 = ε/M , where M = . Then, there exists an N2 ∈ N such that for all n ≥ N2 ,
|b|2
(2.2) |b n − b| < ε/M .

Let N = max{N1 , N2 }. Then, by (2.1) and (2.2), for all n ≥ N ,


¯ ¯
¯ 1
¯ − 1 ¯ < M |b n − b| < ε.
¯
¯b b¯
n
Since ε > 0 was arbitrary, we are done.

Remark. The converse statements generally do not hold in the above theorem, mostly because
the convergence of sequences such as {a n +b n }, {a n b n } and {a n /b n } may not in general imply the
convergence of the individual sequences {a n } and {b n }.

END OF LECTURE 8

2.2. Infinite Series.

Definition 2.9. A infinite series of real numbers is a formal expression of the form

X
a0 + a1 + a2 + · · · or an .
n=0
P∞
Given a series n=0 a n , its sequence of partial sums (sops) is the sequence {s n }n∈N given by

sn = a0 + · · · + an , n ∈ N.

a n is convergent and has sum ℓ if {s n }n∈N is convergent with limit ℓ. Otherwise,


P∞
We say that n=0
P∞
we say that n=0 a n is divergent.
P∞ 1
Example 10. (a) (Harmonic series) Claim. n=1 n is divergent.
16

Proof. Observe that

s1 = 1
1
s2 = 1 +
2
1 1 1 1 1 1
s4 = 1+ + + ≥ 1+ + +
2 3 4 2 4 4
1 1 1 1 1 1 1 1
s8 = 1+ + + + +···+ ≥ 1+ + +
2 3 4 5 8 2 2 2
..
.
1 1 1 1 1
s 2k ≥ 1+ + 2 + 4 + · · · + 2k−1 k = 1 + k .
2 2 8 2 2
For any m ∈ R, there is some k ∈ N such that s 2k > m. Thus, {s n } is divergent. □
P∞ 1
(b) Claim. n=1 n 2 is convergent.

Proof. Observe that


n
X 1 Xn 1 1 1
sn < 1 + = 1+ − = 1 + 1 − → 2 as n → ∞.
k=2 k(k − 1) k=2 k − 1 k n

Remark. We have seen an example of a telescoping series above. These are series of the form
P∞
c n − c n+1 for some sequence {c n }. Note that, in this case, s n = c 1 − c n+1 , so the convergence of
n=1
the series entirely depends on the convergence of {c n }.
P∞ n 1
(c) (Geometric) Claim. Let −1 < x < 1. Then, n=0 x converges and its sum is 1−x
. For |x| ≥ 1,
P∞ n 1
n=1 x = 1−x
diverges.

Proof. Observe that for x ̸= 1,


1 − x n+1 1 x
sn = 1 + x + · · · + x n = = − xn.
1−x 1−x 1−x
We inspect the behavior of {x n } as n → ∞. In the case when |x| < 1, we use the fact that
(1 + y)n ≥ n y for any y > 0 and n ∈ N, to observe that
µ ¶
n 1
(1 + (1/|x|) − 1) > n − 1 = nc.
|x|
Thus, |x|n < 1/(nc). By the “squeeze theorem" stated in HW04, we have that limn→∞ x n = 0.
17

In the case, when |x| > 1, (1 + |x| − 1)n > n(|x| − 1). Thus, for any R ∈ R, by the Archimedean
principle, there exists an N ∈ N such that |x|N > R. Thus, {x n } is unbounded, and therefore,
divergent.
P∞ n
Returning to the series n=0 x . When |x| < 1, we use the limit laws of convergent sequences
to say that
1
. lim s n =
1−x n→∞

When |x| > 1, x =


̸ 1, we use the fact that the sum of a convergent and divergent sequence is, in
fact, divergent. Thus, {s n } is divergent.
P∞ n
For x = 1, observe that s n = n + 1 → ∞ as n → ∞. Thus, n=0 (1) diverges. □
P
Theorem 2.10. Let a n be a convergent series. Then, limn→∞ a n = 0.

Proof. Let ℓ = a n . Let ε > 0. Then, there exists N ∈ N such that |s n − ℓ| < ε/2 for all n ≥ N .
P

Observe that a n = S n − S n−1 , n ≥ 1. Thus, |a n | ≤ |S n − ℓ| + |S n−1 − ℓ| < ε for all n ≥ N + 1. □

END OF LECTURE 9
We will work with series with non-negative terms. Note that the sops for such series are
monotonically increasing, so the task of showing convergence reduces to showing the (upper)
boundedness of the sops.

Theorem 2.11 (Comparison Test). Suppose there exist constants M ∈ N and C > 0 such that

0 ≤ an ≤ C bn ∀n ≥ M .
P P P P
Then, a n if b n is convergent. In other words, b n diverges if a n diverges.
P P
Proof. Let {s n } and {t n } denote the sequence of partial sums of a n and b n respectively. Since
P
b n converges, there exists an N > M and an L > 0 such that t n ≤ L for all n ≥ N . Now,

sn ≤ C tn ∀n ≥ N .

Thus, {s n } — being a bounded monotone sequence — converges. □


P∞ 1
Example 11. (a) (p-series) Claim. n=1 diverges for p ≤ 1 and converges for p > 1.
np
Proof. For p ≤ 0, 1/n p → ∞, so the series must diverge.
1
For 0 < p ≤ 1, n p ≤ n. Thus, np
≥ n1 . By the Comparison Test, the series diverges.
1
For p ≥ 2, n p ≥ n 2 . Thus, np ≤ n12 . By the Comparison Test, the series converges.
18
P
For 1 < p < 2, one must directly analyze the s.o.p.s. as we did with the case of 1/n:

s1 = 1
1 1 1
s3 = 1 + + ≤ 1 + 2
2p 3p 2p
1 1 1 1 1 1
s7 = 1+ p + p + p +···+ p ≤ 1+2 p +4 p
2 3 4 7 2 4
..
.
1 1 1
s 2k −1 ≤ 1 + 2 p
+ · · · + 2k−1 p(k−1) < .
2 2 1 − 2p−1
Exercise: complete the proof. □

Many other tests are derived from the comparison test. We will note two here, but prove only
one.
P
Theorem 2.12 (Ratio Test). Let a n be a series of non-negative terms. Suppose
a n+1
lim = L.
n→∞ a n

(1) If L < 1, the series converges.


(2) If L > 1, the series diverges.
(3) If L = 1, then the test is inconclusive.

END OF LECTURE 10
Proof. Case 1. L < 1. Choose an r such that L < r < 1. Choosing ε = r − L >0, we obtain an N ∈ N
a n+1
such that < L + ε = r for all n ≥ N . Thus,
an
a N +1 < a N r
a N +2 < a N +1 r < a N r 2
a N +3 < a N +2 r < a N r 3
..
.
a N +k < aN r k .

In other words, for n ≥ N , a n ≤ ar NN r n = cr n . By the Comparison Test, and the convergence of the
geometric series r n , r < 1, we have the convergence of a n .
P P

Case 2. L > 1. Choose R such that 1 < R < L. Then, choosing ε = L − R, we have that for some
a n+1
N ∈ N, > L − ε = R > 1 for all n ≥ N . Thus, a n+1 > a n for all n ≥ N . The sequence {a n } cannot
an P
converge to 0. Thus, a n diverges.
19

P 1 n +1 P1
Case 3. L = 1. converges while limn→∞ = 1. On the other hand, diverges while
n2 n n
n2 + 1
limn→∞ = 1. □
n2
P
Theorem 2.13 (Root Test). Let a n be a series of non-negative terms. Suppose

p
n
lim a n = R.
n→∞

(1) If R < 1, the series converges.


(2) If R > 1, the series diverges.
(3) If R = 1, then the test is inconclusive.

Proof. Exercise!
1/n 2 . How does one compute limn→ n 1/n without using
P P
For Case (3), we can take 1/n and
Fms? □
P P
Theorem 2.14 (Limit Laws for Series). Suppose a n and b n converge with sums a and b respec-
tively. Then, for constants ℓ and m, ℓa n + mb n converges to l a + mb. Suppose |a n | and |b n |
P P P

converge. Then, so does |ℓa n + mb n | for any choice of ℓ and m in R.


P

Proof. Exercies! □

Corollary 2.15. Suppose a n converges and b n diverges. Let m ∈ R \ {0}. Then, (a n + b n )


P P P
P
diverges, and mb n diverges.
P P
Definition 2.16. A series a n of real numbers is said to converge absolutely if |a n | converges. A
P P P
series a n of real numbers is said to converge conditionally if |a n | diverges but a n converges.
P P P
Theorem 2.17. If a n converges absolutely, it must converge. Moreover, | an | ≤ an

Proof. We will construct a new series as follows:

b n = a n + |a n |.

P
Observe that 0 ≤ b n ≤ 2|a n |. Thus, by the comparison test, b n converges. Now, by the limit laws
P P
for convergent series, a n = (b n − |a n |) converges. □
P (−1)n
Example 12. Claim. n is convergent.
20

Proof. Note that

s 1 = −1
1 1
s 3 = −1 + − > s1
2 3
1 1 1 1
s5 = −1 + − + − > s 3
2 3 4 5
..
.
µ ¶ µ ¶ µ ¶
1 1 1 1
s 2k+1 = −1 + + − + + · · · − < 0.
2 3 4 2k + 1
Thus, {s 2k+1 } being a bounded increasing sequence, converges to some limit, say ℓ.

1
s2 = −
2
1 1 1
s 4 = −1 + − + < s 2
2 3 4
..
.
µ ¶ µ ¶ µ ¶
1 1 1 1 1
s 2k = −1 + − + − + · · · + ≥ −1.
2 3 4 5 2k
Thus, {s 2k } being a bounded decreasing sequence, converges to some limit, say m. Moreover,
1
s 2k+1 = s 2k + 2k+1 . So, by limit laws for sequences, ℓ = m. Exercise: why does this suffice to claim
that {s n } converges? □

Theorem 2.18 (Alternating Series Test/Leibniz Test). Suppose {a n } is an decreasing sequence of


positive numbers going to 0. Then, (−1)n a n converges. Denoting the sum by S, we have that
P

0 < (−1)n (S − s n ) < a n+1 .


P (−1)n
Proof. Same principle as the example of n . □

Remark. The estimate in AST allows us to estimate sums of alternating series within any pre-
(−1)n
scribed error. For instance, to know ∞
P
n=1 n up to an error of 0.01. I need to find n so that
1
|S − s n | < .
100
Take n = 99, or the sum of the first 99 terms.

END OF LECTURE 11
21

3. L IMITS AND CONTINUITY

3.1. Limit of a function.

Definition 3.1. Given a real number p and an ε > 0, the ε-neighborhood of p is the open interval

Nε (p) = (p − ε, p + ε) = {x ∈ R : p − ε < x < p + ε} = {x ∈ R : |x − p| < ε}.

Definition 3.2. (a) Given a function f that is defined on some I = (a, p)∪(p, b), a < b, we say that
f has a limit L as x approaches p if:
for every ε > 0, there is a δ > 0 such that whenever 0 < |x − p| < δ, we have that | f (x) − L| < ε. OR
for every ε > 0, there is a δ > 0 such that whenever x ∈ Nδ (p) \ {p}, f (x) ∈ Nε (L).
This is denoted by
lim f (x) = L.
x→p
——————–Skipped one-sided limits!!——————
Given a function f : (a, p) → R we say that f has a left-hand limit L as x approaches p if:
for every ε > 0, there is a δ > 0 such that whenever p − δ < x < p, we have that | f (x) − L| < ε.
This is denoted by
lim f (x) = L.
x→p −
Given a function f : (p, b) → R we say that f has a right-hand limit L as x approaches p if:
for every ε > 0, there is a δ > 0 such that whenever p < x < p + δ, we have that | f (x) − L| < ε.
This is denoted by
lim f (x) = L.
x→p +
——————————————————————

Remark. Check that limx→p f (x) = L is equivalent to limh→0 f (x + h) = L.

Theorem 3.3 (Uniqueness of limits). Suppose f has limits L 1 and L 2 at p. Then, L 1 = L 2 .

Proof. Similar to the proof of uniqueness of sequential limits. □

Example 13. (1) (Constant functions) f (x) = c, for some fixed c ∈ R. Let p ∈ R. Let ε > 0. Let δ = 1
(or any ohter positive number!). Then, whenever 0 < |x − p| < 1, we have that | f (x) − c| = 0 < ε.
Thus, for every p ∈ R,
lim f (x) = c.
x→p
(2) (Identity function) f (x) = x. Fix p ∈ R. Let ε > 0. Choose δ = ε. Then, whenever |x − p| < δ,
we have that | f (x) − f (p)| = |x = p| < δ = ε. Since, ε > 0 and p ∈ R were arbitrary, we have that
limx→p x = p for all p ∈ R.
22
p p
(3) f (x) = x, x > 0. We will show that limx→p f (x) = p, p > 0. Let ε > 0.
p p
Scrapwork. We want to produce a δ > 0 such that if 0 < |x − p| < δ, then | x − p| < ε. Note
that
p p |x − p| 1
| x − p| = p p ≤ p |x − p|.
| x + p| p

END OF LECTURE 12
p
Example 14. Let δ = pε > 0. Then, for 0 < |x − p| < δ,
p p |x − p| 1
(3.1) | x − p| = p p ≤ p |x − p| < ε.
x+ p p
Since ε > 0 was arbitrary, we are done.
(4) f (x) = x1 , x ̸= 0. Then, limx→0 f (x) does not exist. Suppose it did, and was L. Then, for any
ε > 0, there exists a δ > 0 such that whenever 0 < |x| < δ, |1/x − L| < ε. In particular, for any
nonzero x, y ∈ (−δ, δ), we must have that

|1/x − 1/y| < 2ε.

Let ε = 1/2 and δ correspond to this ε. Then, by the Archimedean property of R, there is an
1
N ∈ N such that N δ > 1, or N
< δ. Choose, x = N and y = N + 1. Then. 0 < |x|, |y| < δ, but
|1/x − 1/y| = |N + 1 − N | = 1 > 1/2.

Theorem 3.4 (Limit laws for functions). Let f and g be functions such that limx→p f (x) = A and
limx→p g (x) = B . Then,
(1) limx→p ( f (x) ± g (x)) = A ± B .
(2) limx→p ( f (x)g (x)) = AB .
(3) limx→p ( f (x)/g (x)) = A/B , if B ̸= 0.

Proof. We will prove (2). Let ε > 0.


Scrapwork. Note that

| f (x)g (x) − AB | = | f (x)g (x) − f (x)B + f (x)B − AB | ≤ | f (x)||g (x) − B | + |B || f (x) − A|.

ε
Let ε1 = . Let δ1 > 0 be such that whenever 0 < |x − p| < δ1 , then | f (x) − A| < ε1 . In
(2|B | + 1)
particular,
| f (x)| < |A| + ε1 = M .
ε
Let ε2 = . Let δ2 > 0 be such that whenever 0 < |x − p| < δ2 , then |g (x) − A| < ε2 .
2M
23

Finally, let δ = min{δ1 , δ2 }. When 0 < |x − p| < δ,

| f (x)g (x) − AB | = | f (x)g (x) − f (x)B + f (x)B − AB | ≤ | f (x)||g (x) − B | + |B || f (x) − A|


≤ M ε2 + |B |ε1 < ε/2 + ε/2 = ε.

END OF LECTURE 13

3.2. Continuity.

Definition 3.5. Let S ⊂ R, $ f : S → R and p ∈ S. We say that f is continuous at p if: for every ε > 0,
there is a δ > 0 such that whenever |x − p| < δ and x ∈ S, we have that | f (x) − f (p)| < ε, in other
words
lim f (x) = f (p).
x→p

We say that f is continuous on S if f is continuous at each p ∈ S.

Theorem 3.6 (Algebraic combinations of continuous functions). Let f , g be functions that are
continuous at p. Then, f ± g , f g and f /g (when g (p) ̸= 0) are continuous at p.

Example 15. (a) Based on our previous computations, constant functions and the identity func-
tion ( f (x) = x) are continuous on all of R. Using the above theorem, every polynomial, i.e., func-
tion of the form
p(x) = a n x n + · · · a 0 ,
where a 0 , ..., a n are constants, is continuous. Every rational function, i.e., function of the form
p(x)
r (x) = ,
q(x)
where p and q are polynomials, is continuous on the set {x ∈ R : p(x) ̸= 0 and q(x) ̸= 0}. Note.
r (x) may be extended to a continuous function on certain points x, where p(x) = q(x) = 0, but a
priori, the function r (x) itself is considered undefined on such points! E.g., the function
x2 − 1
r (x) =
x −1
is only defined and continuous on R \ {1}. The function

 x 2 −1 , x ̸= 1,
R(x) = x−1
2, x = 1.

is defined on all of R, and is in fact also continuous everywhere (but this requires proof at x = 1).
24

(b) f (x) = x r for any r ≥ 0 is continuous on (0, ∞). In HW 5, you will show that limx→p x n = p n for
every n ∈ N and p ∈ R. Now, let q > 0. There exists a natural number n such that q ≤ n. Thus,

0 ≤ y q − 1 ≤ y n − 1, y >1
0 ≤ 1 − yq ≤ 1 − yn, y ≤ 1.

By the squeezing principle (HW05), limx→1 x q = 1. Now, for any a > 0, note that

|x q − a q | = a q ¯(x/a)q − 1¯ .
¯ ¯

(c) Trigonometric functions will have the meaning (in terms of triangles) that they did in school
(angles are measured in radians!). You may assume all the trigonometric formulas from school.
You will establish their continuity in HW5.

END OF LECTURE 14
Theorem 3.7 (Compositions). Given f : A → R and g : B → R, assume that ran( f ) ⊂ B . Then,
g ◦ f : A → R given by g ◦ f (x) = g ( f (x)) is well-defined. If f is continuous at p ∈ A, and g is
continuous at q = f (p), then g ◦ f is continuous at p.

Proof. Let ε > 0. We must produce a δ > 0 so that whenever |x − p| < δ and x ∈ A, we have that

|g ◦ f (x) − g ◦ f (p)| < ε.

By the continuity of g at q = f (p), there is a τ such that whenever |y − q| < τ and y ∈ B , then

(3.2) |g (y) − g (q)| < ε.

Now, set ε2 = τ. By the continuity of f at p, there is a δ > 0 such that whenever |x − p| < δ and
x ∈ A, we have that
| f (x) − f (p)| < τ.
Thus, whenever |x − p| < δ and x ∈ A, we have that | f (x) − q| < τ and f (x) ∈ B . Thus, by (3.2),

|g ( f (x)) − g ( f (x))| < ε.

Before we proceed to the next theorem about continuous functions, let us prove a result about
sequences.

Lemma 3.8 (Comparison for sequences). Suppose {a n }, {b n } ⊂ R are convergent sequences such
that a n ≤ b n , then L 1 = limn→∞ a n ≤ limn→∞ b n = L 2 .
25

Proof. Suppose not. Say that L 1 > L 2 . Let ε = (L 1 − L 2 )/2. Then, there exists an N1 ∈ N and N2 ∈ N
such that L 1 − ε < a n and b n < L 2 + ε for all n ≥ max{N1 , N2 }. I.e., b n < (L 1 + L 2 )/2 < a n for all
n ≥ max{N1 , N2 }. This is a contradiction. □

A common test used to detect zeros of a polynomial are to look for places where the polyno-
mial “changes sign". This comes from an important property of continuous functions!

Theorem 3.9 (The Intermediate Value Theorem). Let f : [a, b] → R be a continuous function.
Suppose y is a real number lying between f (a) and f (b). Then, there is a c ∈ [a, b] such that f (c) =
y.

We will assume that f (a) ≤ y ≤ f (b). The case f (b) ≤ y ≤ f (a) can be handled similarly.

Proof. f y = f (a) or y = f (b), then c = a or c = b, respectively, works. So, we may assume that
f (a) < y < f (b). Let
S = {x ∈ [a, b] : f (x) < y} ⊂ [a, b].

Since a ∈ S, S is nonempty. Since b ≥ x for all x ∈ S, S is bounded above. Thus,

c := sup S exists.

We first show that f (c) ≤ y. Let n ∈ P. Since c − 1/n is not an upper bound of S, there is an x n ∈ S
such that
1
c−
< x n ≤ c.
n
Thus, by the above lemma, limn→∞ x n = c. By the above lemma and sequential characterization
of continuity, y ≥ limn→∞ f (x n ) = f (c).
Next, we show that y ≤ f (c). Since b is an upper bound of S, c ≤ b. Moreover, f (c) ≤ y < f (b),
so c ̸= b. Let N ∈ N such that c + 1/n < b for all n ≥ N . Then, f (c + 1/n) ≥ y. By the sequential
char. of cont. and the above lemma, f (c) = limn→ f (c + 1/n) ≥ y. □

Corollary 3.10 (Bolzano’s theorem). Let f : [a, b] → R be a continuous function such that f (a)
and f (b) assume opposite signs (so, are non-zero). Then, there is at least one c ∈ (a, b) such that
f (c) = 0.

Informal discussion. A continuous function on a circle is the same as a continuous 2π-


periodic function on the real line. For such a function f , let

g (θ) = f (θ) − f (θ + π).


26

Then, g (0) = f (0) − f (π) = −( f (π) − f (2π)) = −g (π). Either g (0) = 0, or g takes opposite signs at
0 and π. In either case: f takes the same values on a pair of anti-podal points. This is the 1-
dimensional case of the Borsuk–Ulam theorem. This is the basis of statements such as, "At any
given time, there are two polar opposite points on the Earth’s equator where the temperature is
exactly the same!"
END OF LECTURE 15
————Skip!————–

Theorem 3.11 (Existence of nth roots). Given a positive integer n and a positive number a, there
is exactly one positive nth root of a.

Proof. Let c > 1 such that a ∈ [0, c]. Let f (x) = x n . Then, f (0) = 0 and f (c) = c n > c > a > 0. By IVT,
there is some b ∈ (0, c) such that b n = a. Now, since 0 < x < y ⇒ x n < y n , we are done. □

————Skip!————–

Definition 3.12. A function f : S → R is said to be bounded above on S if there is some U ∈ R such


that
f (x) ≤ U ∀x ∈ S.
A function f : S → R is said to be bounded below on S if there is some L ∈ R such that

f (x) ≥ L ∀x ∈ S.

A function f : S → R is said to be bounded on S if there is some M > 0 such that

| f (x)| ≤ M ∀x ∈ S.

Theorem 3.13. A function f : S → R is bounded if and only if it is bounded above and bounded
below on S.

Example 16. The function 1/x, x ̸= 0, is neither bounded above nor bounded below on its do-
main. However, it is bounded below on (0, 1).

Theorem 3.14. Let f : [a, b] → R be a continuous function. Then f is bounded on [a, b].

Proof. Given any closed interval I = [c, d ] with midpoint e, if f is bounded on I − = [c, e] and I + =
[e, d ], then f is bounded on [c, d ]. Suppose f is not bounded on [a, b]. Let I 0 = [a, b] and a 0 = a.
Then, by the contra-positive of the above statement, f is either not bounded on I 0− = [a, (a +b)/2]
or on I 0+ = [(a + b)/2, b]. Pick the one on which it is not bounded (pick left to break a tie), and call
it I 1 . Call its left endpoint a 1 . Continuing this way, inductively, we obtain a sequence of intervals
27

+ −
I 0 = [a 0 , b 0 ],...,I k = [a k , b k ],..., so that f is unbounded on I k and I k is either I k−1 or I k−1 . Since we
are halving the intervals at each stage, |b j − a j | < (b − a)/2 j .
Let s = sup{a j : j ∈ N}. Justify why s exists, and show that s ∈ [a, b].
By continuity, there is a δ > 0 such that if x ∈ Nδ (s) ∩ [a, b], then | f (x) − f (s)| < 1. I.e., | f (x)| <
1 + | f (s)| on x ∈ Nδ (s) ∩ [a, b]. Let N ∈ N such that for all n ≥ N , a n ∈ Nδ/2 (s) ∩ [a, b]. This works
because of the monotonicity of {a n }. Choose N ′ ≥ N such that for all n ≥ N ′ , (b − a)/2n < δ/2.
Thus,
|b n − s| < |b n − a n | + |a n − s| < δ ∀n ≥ N ′ .
This contradicts the fact that f is unbounded on these intervals! □
END OF LECTURE 16
Definition 3.15. A function f : S → R is said to have a global maximum (minimum) at p ∈ S if
f (p) ≥ (≤) f (x) for all x ∈ S.

Theorem 3.16 (Extreme value theorem). Let f : [a, b] → R be a continuous function. Then, f
attains a global maximum and a global minimum in [a, b].

Proof. Let s = sup{ f (x) : x ∈ [a, b]}. We show that there is some c ∈ [a, b] such that f (c) = s.
Suppose not. Let g (x) = s − f (x). Then, g is continuous and g (x) > 0 on [a, b]. Thus, 1/g (x) is
bounded on [a, b]. Thus, by the above theorem, there is an M > 0 such that 1/g (x) ≤ M for all
x ∈ [a, b]. Thus, s − f (x) ≥ 1/M for all x ∈ [a, b]. Thus, s − (1/M ) ≥ f (x) on [a, b]. This contradicts
the fact that s is the least upper bound.
The first part of the argument applied to − f on [a, b] shows that there exists a d ∈ [a, b] such
that − f (d ) ≥ − f (x) for all x ∈ [a, b]. Thus, f (x) ≤ f (d ) on [a, b]. □

Corollary 3.17. Let f : [a, b] → R be a continuous function. Then, f ([a, b]) = [min[a,b] f (x), max[a,b] f (x)].

Theorem 3.18 (Composition for limits, based on HW05). Let f and g be functions such that

lim f (x) = L and lim g (y) = M .


x→p y→L

Moreover, suppose that for some δ > 0, if 0 < |x − p| < δ, then | f (x) − L| > 0. Then,

lim g ( f (x)) = lim g (y).


x→p y→L
28

4. D IFFERENTIATION

Fun fact: continuous functions can be very weird. They can have “corners" or “be jagged" at
every point!

Definition 4.1. Given a function f : (a, b) → R and p ∈ (a, b) we say that f is differentiable at p if
the limit
f (p + h) − f (p) f (x) − f (p)
lim = lim
h→0 h x→p x −p
′ ′
exists, in which case, we call the above limit f (p). f (p) is also sometimes known as the rate of
change of f at p. If f is differentiable at each p ∈ (a, b), it is said to be differentiable on (a, b), and
f ′ : (a, b) → R is the derivative function of f .

Example 17. (a) (Constant functions) f (x) = c, c ∈ R, is differentiable on R and f ′ (x) ≡ 0. For any
x ∈ R and h ̸= 0,
f (x + h) − f (x)
= 0.
h
(b) (Linear functions) f (x) = ax +b, x, a, b ∈ R, is differentiable at each x ∈ R, and f ′ (x) = a. Proof:
homework!
(c) (Natural powers) f (x) = x n , n ∈ P. We will use the formula (telescoping sum)
n−1
a n − b n = (a − b) a k b n−1−k .
X
k=0
Note that, for any x ∈ R and n ∈ P,
(x + h)n − x n n−1
(x + h)k x n−1−k .
X
=
h k=0

But polynomials are continuous, so limh→0 k=0 (x + h)k x n−1−k = n−1


Pn−1
xn − 1 = nx n−1 .
P
k=0
(d ) (Sine and cosine) Homework! Uses addition formulas, and limx→0 sinx x = 1.

Theorem 4.2. Let f : (a, b) → R be a function that is differentiable at p ∈ (a, b). Then, f is contin-
uous at p.

Proof. x ∈ (a, b) \ {p}, we may write


f (x) − f (p)
f (x) − f (p) = (x − p) .
x −p
By the limit laws, the limit of the R.H.S., as x → p, exists, and is equal to 0. Thus,

lim
x→ f (x)= f (p).


29

Remark. The converse need not be true. Consider f (x) = |x|. The function is continuous (why?),
but observe that 
f (h) − f (0) 1, h > 0,
=
h −1, h < 0.
Thus, taking {(−1)n n −1 }, we see that the derivative at 0 does not exist.

Theorem 4.3 (Algebra of derivatives). Let f and g be defined on (a, b) and differentiable at p ∈
(a, b). Then,
(1) so is f + g , and ( f + g )′ (p) = f ′ (p) + g ′ (p),
(2) (product rule) so is f g , and ( f g )′ (p) = f (p)g ′ (p) + f ′ (p)g (p),
g (p) f ′ (p)− f (p)g ′ (p)
(3) (quotient rule) if g (p) ̸= 0, so is f /g , and ( f /g )′ (p) = g 2 (p)
.

Proof. We only prove the quotient rule in the special case of f ≡ 1. Note that
1 1
g (p+h)
− g (p) g (p + h) − g (p) 1 1
=− .
h h g (p) g (p + h)
By the continuity of g at p and the algebra of limits of functions,
1 1
g (p+h)
− g (p) g (p + h) − g (p) 1 1 g ′ (p)
lim = − lim =− .
h→0 h h→0 h g (p) g (p + h) g (p)2

Example 18. We now have the differentiability of polynomials and rational functions in their
domains of definition.

END OF LECTURE 18
Definition 4.4. Let f : A → B be a one-to-one and onto function. f is said to be invertible on A,
and given, any y ∈ B , there is a unique x y ∈ A such that f (x y ) = y. Define the inverse of f function
f −1 : B → A as
f −1 (y) = x y .
Then, f −1 ( f (x)) = x for all x ∈ A, and f −1 ( f (y)) = y for all y ∈ B .

Theorem 4.5 (Monotone/Continuous/Differentiable Inverse Function Theorem). Let f : [a, b] →


R be an invertible function with range J . Then
(i ) If f is strictly increasing (decreasing), so is f −1 : J → R.
(i i ) If f is continuous on [a, b], then f is strictly monotone on [a, b], and f −1 is continuous on
J.
30

(i i i ) If f is continuous on [a, b], f is differentiable at p ∈ (a, b) and f ′ (p) ̸= 0, then f −1 is differ-


entiable at q = f (p), and
¢′ 1
f −1 ( f (p)) =
¡
′ (p)
.
f
Proof. (i ) Almost follows from definition, so left as an exercise.
(i i ) First, we show the strict monotonicity of f .
Case 1. f (a) < f (b). Let g : [a, b] → R be given by g (y) = f (y) − f (a). Note that g is cont. on [a, b]
and g (b) > 0. Suppose g (y) ≤ 0 for some y ∈ (a, b). Then g vanishes at some c ∈ [y, b) by IVT, and
f (c) = f (b). This contradicts the one-t-one-ness of f . So, f (a) < f (y) for all y ∈ (a, b]. Similarly,
by considering h(y) = f (y) − f (b) on [a, b], we have that f (y) < f (b) for all y ∈ [a, b).
Thus,

(4.1) f (a) < f (y) < f (b), if a < y < b.

Now, let x, y ∈ [a, b] such that x < y. If x = a, we already have that f (x) < f (y) by (4.1). Now,
suppose a < x < y. Then, applying the above argument to f restricted to the domain [a, y], we
have that f (a) < f (x) < f (y). Thus, f is strictly increasing on [a, b]
Case 2. f (a) > f (b). Apply Case 1 to − f to obtain that f is strictly decreasing. □

END OF LECTURE 19

Proof. (i i ) We have already shown that f is strictly monotone. Assume, WLOG (why?), that f is
strictly increasing. Then, by IVT, we have that J = [ f (a), f (b)].
Case 1. p ∈ (a, b). Let q = f (p). Let d = 21 min{p − a, b − p} > 0. Let ε > 0. Suppose ε ≤ d . Then

a < p − ε < p < p + ε < b.

Thus, since f is strictly increasing, we have that

f (a) < f (p − ε) < q < f (p + ε) < f (b).

Let δε = min{ f (p + ε) − q, q − f (p − ε)}. Now, whenever |y − q| < δ, we have that

f (p − ε) < q − ε < y < q + ε < f (p + ε).

Thus,
p − ε < f −1 (y) < p + ε.

If ε > d , choose δε as δd . Then, whenever |y − q| < δd , we have that | f −1 (y) − f −1 (q)| < d < ε.
Case 1. p = a or p = b. Homework!
31

(i i i ) We prove the differentiability of f −1 at q = f (p). Given k such that q + k ∈ J , Let h(k) =


f −1 (q + k) − f −1 (q). Then, h(k) ̸= 0 whenever k ̸= 0, (since f −1 is is one-to-one). Then h(k) + p =
f −1 (q + k), or k = f (h(k) + p) − f (p). Thus,

f −1 (q + k) − f −1 (q) h(k) 1
= = f (p+h(k))− f (p)
.
k f (p + h(k)) − f (p)
h(k)
By continuity of f −1 , limk→0 h(k) = 0. Since h(k) ̸= 0, whenever k ̸= 0, we have by the composition
rule and the limit laws that
f −1 (y + k) − f −1 (y) 1 1 1
lim = lim f (x+h(k))− f (x)
= f (x+h)− f (x)
= ′ (x)
.
k→0 k k→0 limh→0 f
h(k) h

Example 19. The above theorem immediately gives the continuity and differentiability of f (x) =
1
x 1/n for x > 0 and n ∈ P. Moreover, f ′ (x) = nx (n−1)/n .
By the algebra of derivatives and induction (or the chain rule — which we haven’t done yet!),
¢p
one obtains the continuity and differentiability of f (x) = x 1/q , x > 0, for each p ∈ Z and q ∈ P.
¡
p
Moreover, f ′ (x) = q x (p−q)/q .

Example 20. We may now define inverse trigonometric functions. As convention, we invert
the sine function on [−π/2, π/2], the cosine function over [0, π] and the tangent function over
(−π/2, π/2).

END OF LECTURE 20

Theorem 4.6. Let f : (a, b) → R and g : (c, d ) → R such that f ((a, b)) ⊂ (c, d ). Say f is differentiable
at p and g is differentiable at q = f (p). Then, g ◦ f is differentiable at p, and

(g ◦ f )′ (p) = g ′ ( f (p)) · f ′ (p).

Proof. Let h = g ◦ f . Then, the relevant difference quotient is


h(p + k) − h(p) g ( f (p + k)) − g ( f (p))
= .
k k
Let ℓ(k) = f (p + k) − f (p). Note that ℓ is a continuous function of k. We have that
g ( f (p + k)) − g ( f (p)) g (q + ℓ(k)) − g (q) f (p + k) − f (p)
= ,
k f (p + k) − f (p) k
where the last equality holds only if f (p + k) ̸= f (p). Idea: what if we could directly write g (q +
ℓ(k)) − g (q) directly as ℓ(k)G(k), where G(k) admits a limit as k → 0?
32

Define 
 g (q+ℓ)−g (q) , ℓ ̸= 0, q + ℓ ∈ (c, d )

G(ℓ) =
g ′ (q), ℓ = 0.
Since limℓ→0 G(ℓ) = G(0), G is continuous at ℓ = 0. Note that

g (q + ℓ) − g (q) = ℓG(ℓ).

Thus,
h(p + k) − h(p) g (q + ℓ) − g (q) ℓ(k)G(ℓ(k))
= = .
k k k
Now, limk→0 ℓ(k)
k
= f ′ (p), by assumption, and limk→0 G(ℓ(k)) = G(0) = g ′ (q) by the continuity of ℓ
and G at 0. Thus, we are done. □

We will now see some applications of derivatives.

Definition 4.7. Let f : A → R. We say that f attains a local maximum (minimum) at a ∈ A if there
exists a δa > 0 such that f (a) ≥ f (x) ( f (a) ≤ f (x)) for all x ∈ A ∩ Nδa (a).

Theorem 4.8 (Extreme=>Critical). Let f : (a, b) → R. Let c ∈ (a, b) such that f is differentiable at c
and f attains a local extremum at c. Then, f ′ (c) = 0.

Remarks. All the conditions are very important!


(1) If the extremum occurs at one of the endpoints of an interval, the function may have a
nonzero derivative there. E.g. f (x) = x when restricted to [0, 1] has local max and min on
[0, 1] at 1 and 0 resp.
(2) f need not even be differentiable at points where local extrema are defined. E.g. f (x) = |x|
at x = 0.
(3) The converse of the above theorem is not true. Take f (x) = x 3 . f does not attain a local
extremum at x = 0, and yet the derivative vanishes there.

END OF LECTURE 21

Proof. We assume (WLOG) that f attains a local maximum at c. Then, there is a δ > 0 such that
f (c) ≥ f (x) for all x ∈ Nδ (c) ∩ (a, b). We may further assume that Nδ (c) ⊂ (a, b).
By the sequential characterization of limits, we have that for any sequence {c n } ⊂ (a, b) such
that limn→∞ c n = c, we have that
f (c n ) − f (c)
lim = f ′ (c).
n→∞ cn − c
33

Now, for c − δ < x < c, we have that


f (x) − f (c)
≥ 0.
x −c
Let N ∈ N be such that c −δ < c −1/n for all n ≥ N . Then, by the comparison lemma for sequences,
we have that
f (c − 1/n) − f (c)
lim = f ′ (c) ≥ 0.
n→∞ 1/n
On the other hand, for c < x < c + δ, we have that
f (x) − f (c)
≤ 0.
x −c
Thus, using the sequence {c + 1/n}, we have that f ′ (c) ≤ 0. Combining the two bounds, we have
our result. □

The theorem is most helpful in the following way: to identify potential points of local extrema
within open intervals where f is differentiable. Thereafter, one needs to do some local analysis
at the potential points. For the local analysis, one inspects the sign of the derivative ‘just a bit
before’ and ‘just a bit after’ the point.

Example 21. Let f : R → R be given by

f (x) = (x − 1)2 + |x| + x 3 .


p
The potential local extrema are at x = 0 (point of non-differentiability), x = −3 and x = 2−1. We
need some more machinery to say which of these are points of extrema, and of what type.

Theorem 4.9 (Mean Value Theorem). Let f : [a, b] → R be continuous on [a, b] and differentiable
f (b)− f (a)
on (a, b). Then, there is some c ∈ (a, b) such that f ′ (c) = b−a
.

Proof. Special case. f (a) = f (b). This is referred to as Rolle’s theorem. By the extreme value
theorem, there exist c 1 , c 2 ∈ [a, b] where f attains its global extrema on [a, b]. If at least one of c 1
or c 2 are in (a, b), we are done. If f (a) = f (b) = c 1 = c 2 . Then, f is a constant function, in which
case, f ′ ≡ 0 = on [a, b].

END OF LECTURE 22
f (b)− f (a)
General case. We want to apply Rolle’s to a function G such that G ′ (c) = 0 ⇐⇒ f ′ (c)− b−a
= 0.
f (b)− f (a)
Let g (x) = f (x) − x b−a . Then, g is cont. on [a, b] and differentiable on (a, b). Moreover,
g (a) = b f (a) − a f (b) = g (b). Thus, by Rolle’s theorem, there is some c ∈ (a, b) such that g ′ (c) = 0,
i.e.
f (b) − f (a)
f ′ (c) = .
b−a
34


Applications of MVT

Theorem 4.10. Let f : [a, b] → R be continuous on [a, b] and differentiable on (a, b). Then
(1) if f ′ (x) > 0 for all x ∈ (a, b), f is strictly increasing on [a, b],
(2) if f ′ (x) < 0 for all x ∈ (a, b), f is strictly decreasing on [a, b],
(3) if f ′ (x) = 0 for all x ∈ (a, b), f is a constant function on [a, b].

Proof. We prove only (1) and leave the rest as an exercise. Let x, y ∈ [a, b] such that x < y. Apply
the MVT on f restricted to [x, y], we have the existence of some z ∈ (x, y) such that
f (y) − f (x)
0 < f ′ (z) = .
y −x
Thus, f (y) > f (x). □
Remarks. (1) What if f ′ (c) > 0 or f ′ (c) < 0 for some c ∈ (a, b)?
(2) What about the converse of each part above? Remember that x 3 is a strictly increasing func-
tion.
(3) You are allowed to use the correct variations of the above result.

Example 22. The above theorem gives the first derivative test whose formal statement is given
as Theorem 4.8 in Apostol. Let us return to f (x) = (x − 1)2 + |x| + x 3 /3. Note that

x 2 + 2x − 3 = (x − 1)(x + 3), x < 0,

f (x) = ³ p ´³ p ´
 x − ( 2 − 1) x + 2 + 1) , x > 0.

Test at x = −3: f ′ (x) > 0 on (−4, −3) and f ′ (x) < 0 on (−3, 0), so f (x) ≤ f (−3) for all x(−4, 0). So,
−3 is a point of local max.
p
Test at x = 0: f ′ (x) < 0 on (−1, 0), and f ′ (x) < 0 on (0, 2 − 1). So, f (x) > f (0) > f (y) for every
p
x ∈ (−1, 0) and y ∈ (0, 2 − 1). Thus, x = 0 is not a point of local max/min.
p
Test at x = 2 − 1 Homework!

Definition 4.11. Let f : (a, b) → R be a differentiable function such that f ′ : (a, b) → R is also
differentiable. Then, the ( f ′ )′ is denoted by f ′′ and is called the second derivative of f . One may
defined the nth derivative of f inductively, and denote it by f (n) . If f (n) exists for all n ∈ N, we say
that f is infinitely differentiable on (a, b)

Self-study: second derivative test


Many functions we encounter such as polynomials and trigonometric functions are infinitely
differentiable.
35

END OF LECTURE 23

Definition 4.12. If a function f has an nth derivative at x 0 , its nth Taylor polynomial at x 0 is the
polynomial

x f (n) (x 0 )
P n0 (x) = P n (x) = f (x 0 ) + f ′ (x 0 )(x − x 0 ) + f ′′ (x 0 )(x − c)2 + · · · + (x − x 0 )n .
n!
P nc can be used to approximate the function f . It would be useful to under stand what the
error term looks like. There are many known forms of the error term. We mention one.

Theorem 4.13 (Taylor’s theorem). Let f : (a, b) → R be (n + 1)-times differentiable on (a, b). Let
x 0 ∈ (a, b). Then, for any x ∈ (a, b), there exists a c(x) between x and x 0 such that

x f (n+1) (c(x))
f (x) = P n0 (x) + (x − x 0 )n+1 .
(n + 1)!
Idea: Want to apply Rolle’s theorem to a function G such that
x
′ (n+1) f (x) − P n0 (x)
G (c) = 0 ⇐⇒ f (c) = (n + 1)! .
(x − x 0 )n+1
Could try
x
(n) f (x) − P n0 (x)
f (t ) − t (n + 1)! .
(x − x 0 )n+1
But G(y) ̸= G(c). We will skip the proof, but work out an example.
Question from class. What if we wish to approximate f by P n+1 instead of P n ?
Answer. The best we can say is that there is a function h = h x0 : (a, b) → R such that lim h(x) = 0
x→x 0
and
x
(4.2) 0
f (x) = P n+1 (x) + h(x)(x − x 0 )n+1

for x ∈ (a, b). Thus, we don’t get as much information on the form of the error term as we do in
Taylor’s theorem. Now, if f was (n + 2)-times differentiable, then by Taylor’s theorem, h in (4.2)
would become
f (n+2) (c(x))
h(x) = (x − x 0 )
(n + 2)!
Example 23. Let f (x) = cos(x). We compute the 3rd Taylor polynomials of f at c = 0. Note that
f (0) = 1, f ′ (0) = 0 and f ′′ (0) = −1. Then,
1 2
P 3 (x) = 1 − x .
2!
36

By Taylor’s theorem, there is some y between 0 and x such that


1 2 sin(y) 3
cos(x) = 1 − x + x .
2! 3!
But the error term is nonnegative on [−π, π]. Thus,
x2
cos x ≥ 1 −
2
on [−π, π].

On trigonometric functions: read Section 2.5 in Apostol.


END OF LECTURE 24
37

5. I NTEGRATION

Our goal will be to construct a “reasonable" theory of “area under a graph". There are two
equivalent theories of integration that appear in introductory calculus books — that of Riemann
integration and Darboux integration. We will follow Apostol and discuss Darboux integration.
The word “reasonable" above refers to the fact that we want area, which is a function whose
inputs are sets and outputs are real numbers, to satisfy certain rules (axioms) such as positivity,
finite additivity, monotonicity under contaiments, etc. It is possible that for any such theory,
the areas of certain sets just cannot be measured, i.e., the sets are not ‘measurable’! You will
encoutner all of this more formally in a measure theory course, but we will see some indication
of this in this course as well.
we will perform integration on a closed interval, i.e., a set of the form [a, b], a < b.

Definition 5.1. (i) Let a < b. A partition of [a, b] is a finite set

P = {x 0 , ..., x n }

such that a = x 0 < x 1 < · · · < x n−1 < x n = b. The partition P determines n subintervals, [x j −1 , x j ],
1 ≤ j ≤ n, of [a, b]. We refer to [x j −1 , x j ] as the j t h interval of P .

(ii) Given a partition P of [a, b], a refinement of P is a partition P ′ such that, as sets, P ⊆ P ′ ,
i.e., every subinterval determined by P ′ is a subset of some subinterval determined by P .

(iii) Given two partitions P = {x 0 < x 1 < · · · < x n } and Q = {y 0 < y 1 < · · · < y m } of [a, b], their
common refinement is the partition

R = P ∪ Q = {z 0 < z 1 < · · · < z N }.

Remark. It is a (tedious to prove) fact that for any ℓ ∈ {1, ..., N } there is a unique j (ℓ) ∈ {1, ..., n}
and unique k(ℓ) ∈ {1, ..., m} such that

(z ℓ−1 , z ℓ ) = (x j (ℓ)−1 , x j (ℓ) ) ∩ (y k(ℓ)−1 , y k(ℓ) ).

Definition 5.2. A function f : [a, b] → R is called a step function if there is a partition P = {x 0 <
· · · < x n } such that s is constant on each open subinterval of P , i.e., for each 1 ≤ k ≤ n, there is a
real number s k such that
s(x) = s k for x ∈ (x k−1 , x k ).
At the end points of these intervals, the function may be defined as anything.

Example 24. The floor and ceiling functions restricted to closed and bounded intervals. Inifnitely
many steps are not allowed in the above definition.
38

Remark. The sum, difference, product and quotient (when defined) of two step functions, with
partitions P and Q is also a step function with partition P ∪ Q. We demonstrate this with a few
pictures, since the proof is long and boring.

Definition 5.3. Let s : [a, b] → R be a step function with partition P = {x 0 < ... < x n }. Its integral is
Rb
denoted by a s(x)d x and is given by
n
X
s k (x k − x k−1 ).
k=1
Ra Rb
We also define b s(x)d x to be − a s(x)d x.

END OF LECTURE 26

Remark. The partition of a step funciton is not uniquely determined. A refinement of P would
also work, but it would leave the integral unaffected.

Theorems 1.2-1.8 state certain key properties of the above definitions such as
Z b Z b Z b
(c 1 s(x) + c 2 t (x))d x = c 1 s(x)d x + c 2 t (x)d x,
a a a
Z b Z b
s≤t ⇒ s(x)d x ≤ t (x)d x,
a a
and
Z kb Z b
s(x/k)d x = k s(x)d x.
ka a
These can formally proved using the hints listed in 1.15. See also Theorem 11.2.16 in Tao.
To integrate a general function, f : [a, b] → R, we want to approximate it with step functions
(from above and below), i.e., we want to consider two classes of functions

Definition 5.4.

S f = {s : [a, b] → R : s is a step function and s ≤ f },


T f = {t : [a, b] → R : t is a step function and t ≥ f }.

For an unbounded function, such as 1/x on [0, 1], the latter set is empty! So, we stick to
bounded functions.

Lemma 5.5. Let f : [a, b] → R be a bounded function, i.e., there is an M > 0 such that

−M ≤ f (x) ≤ M ∀x ∈ [a, b].


39

Then, sup S f and inf T f exist, and


½Z b ¾ ½Z b ¾
−M (b − a) ≤ sup s(x)d x : s ∈ S f ≤ inf t (x)d x : t ∈ T f ≤ M (b − a).
a a

Proof. Let g (x) = −M , x ∈ [a, b], and h(x) = M , x ∈ [a, b]. Then, g ∈ S f and h ∈ T f . Moreover, g ≤ t
for all t ∈ T f and h ≥ s for all s ∈ S f . Thus,
Z b Z b
−M (b − a) = g (x)d x ≤ t (x)d x, ∀t ∈ T f
a a
Z b Z b
M (b − a) = h(x)d x ≥ s(x)d x, ∀s ∈ S f .
a a

END OF LECTURE 27

Definition 5.6. Let f : [a, b] → R be a bounded function. The lower integral of f is the quantity
½Z b ¾
I ( f ) = sup s(x)d x : s ∈ S f ,
a
and the upper integral of f is the quantity
½Z b ¾
I ( f ) = inf t (x)d x : t ∈ T f .
a

We say that f is Riemann integrable if I ( f ) = I ( f ), in which case this quantity is called the integral
Rb
of f over [a, b] and denoted by a f (x)d x. We set
Z a Z b
f (x)d x = − f (x)d x,
b a
Z a
f (x)d x = 0.
a
Example 25. Show that every step function is integrable and there is no ambiguity in using the
Rb
symbol a s(x)d x for a step function.

Theorem 5.7. Every bounded monotone function on [a, b] is Riemann integrable.

Proof. (WLOG?) Assume f is increasing. Let n ∈ N. Let


b−a
xj = a + j , j = 1, ..., n.
n
Let s, t be step functions on [a, b] given by

s n (x) = f (x j −1 ), x ∈ [x j −1 , x j ),
t n (x) = f (x j ), x ∈ [x j −1 , x j ).
40

Let s n (b) = f (b) = t n (b). Note that


Z b b b−a
Z
t n (x)d x − s n (x)d x = ( f (b) − f (a)) .
a a n
Note Z b Z b
s n (x)d x ≤ I ( f ) ≤ I ( f ) ≤ t n (x)d x.
a a
Letting C = ( f (b) − f (a))(b − a), we have that
C
I(f )− I(f ) ≤ .
n
Taking n → ∞, we obtain the integrability of f . □

Example 26. We claim that for b > 0,


b b p+1
Z
xpd x = .
0 p +1
n
Since x is monotone on [0, b], it is Riemann integrable. Motivated from the previous proof, let
x ℓ = ℓb/n, ℓ = 0, 1, ..., n. Then,
X kb p
b n−1 b Xn µ kb ¶p
µ ¶
p
≤ I (x ) ≤ .
n k=0 n n k=1 n
I.e.,
µ ¶p+1 Ãn−1 ! µ ¶p+1 Ã n !
b X p p b X p
k ≤ I (x ) ≤ k .
n k=0 n k=1
Now, we use that
1p + · · · + (n − 1)p 1
lim = .
n→∞ n p+1 p +1
This follows from the inequalities
n p+1
1p + · · · + (n − 1)p < < 1p + · · · + n p ,
p +1
which can be proved by induction (on n).

END OF LECTURE 28
We will take the following properties of Riemann integration for granted: linearity with re-
spect to integrand (Thm. 1.16); additivity with respect to interval of integration (Thm 1.17); in-
variance under translation (Thm 1.18); epxansion/contraction of interval of integration (Thm
1.19); comparison (Thm 1.20). Try to prove these for extra practice, or see Section 1.27 for the
proofs.
41

Definition 5.8. A function f : A → R is s.t.b. uniformly continuous on A if, for every ε > 0, there
is a δ > 0 such that whenever |x − y| < δ, x, y ∈ A, we have that | f (x) − f (y)| < ε.

Remarks. (1) Uniform cont ⇒ cont.


(2) If f is u.c. on A, then it must be u.c. on every B ⊆ A.

Example 27. (1) f (x) = x is uniformly continuous on R. Note that δ = ε works for each ε > 0.

(2) f (x) = x 2 is not uniformly continuous on R. We need to produce an ε > 0 such that for every
δ, there exist x δ and y δ such that |x δ − y δ | < δ, but | f (x δ ) − f (y δ )| ≥ ε.
Choose ε = 1. Let δ > 0. Let x δ = δ1 and y δ = δ1 + δ2 . Then, |x δ − y δ | < δ, but

| f (x δ ) − f (y δ )| = 1 + δ2 /4 > 1.

FACT. Every continuous function on a closed and bounded interval is, in fact, uniformly contin-
uous! We will not prove this in class, however, this can be proved using a similar technique as was
used for showing that continuous functions on closed and bounded intervals are bounded.

Theorem 5.9. Let f : [a, b] → R be a continuous function. Then, f is Riemann integrable.

Proof. Since f is continuous on [a, b], it is bounded. Thus, I ( f ) and I ( f ) exist.


Let ε > 0. By the uniform continuity of f on [a, b], there exists a δ > 0 such that whenever
b−a
|x − y| < δ for x, y ∈ [a, b], we have that | f (x) − f (y)| < ε. Let n ∈ N such that n < δ. As done in
previous proofs, choose
b−a
xj = a + j , j = 0, ..., n.
n
Now, f |[x j −1 ,x j ] is a continuous, therefore bounded, function that attains its max. value, M j , at
some u j ∈ [x j −1 , x j ] and min. value, m j , at some l j ∈ [x j −1 , x j ]. Thus, on [x j −1 , x j ],

mj ≤ f ≤ Mj.

Moreover, since |u j − l j | < δ, we have that M j − m j < ε.


Now, set

s n (x) = m j , x ∈ [x j −1 , x j )
s n (b) = f (b).

and

t n (x) = M j , x ∈ [x j −1 , x j )
t n (b) = f (b).
42

Then,
Z b Z b
s(x)d x ≤ I ( f ) ≤ I ( f ) ≤ t (x)d x.
a a
Thus,
b−a Xn
0 ≤ I(f )− I(f ) ≤ M j − m j < ε(b − a).
n j =1
Since ε > 0 was arbitrary, we are done! □
END OF LECTURE 29

Theorem 5.10 (M integrals). Let f : [a, b] → R be continuous. Then, there is some c ∈ [a, b] such
that Z b
f (c)(b − a) = f (x)d x.
a

Proof. Let M = max f and m = min f . Then,


1
Z b
m≤ f (x)d x ≤ M .
b−a a
Thus, by IVT, there is some c ∈ [a, b] such that f (c) attains the claimed value. □

Theorem 5.11. Let a < b. Let f : [a, b] → R be Riemann integrable. Then, so is | f |, and
¯Z b ¯ ¯Z a ¯ Z b
¯ ¯ ¯ ¯
¯
¯ f (x)d x ¯¯ = ¯¯ f (x)d x ¯¯ ≤ | f (x)|d x.
a b a

We will now relate the notions of integrability and differentiability.

Theorem 5.12 (The first Fundamental Theorem of Calculus). Let f : [a, b] → R be a Riemann
integrable function. Let F : [a, b] → R be the function given by
Z x
F (x) = f (t )d t .
a
Then, F is a continuous function on [a, b]. Moreover, if f is continuous at p ∈ (a, b), then F is
differentiable at p, and F ′ (p) = f (p).

Proof. Since f is bounded, there exists an M > 0 such that −M ≤ f (x) ≤ M for all x ∈ [a, b]. Now,
let x, y ∈ [a, b] such that x < y. Then,
Z y
F (y) − F (x) = f (t )d t ≤ M (y − x).
x
Similarly, −M (y − x) ≤ F (y) − F (x). Thus,

|F (y) − F (x)| ≤ M |y − x|.


43

Interchanging the role of x and y, we get the same conclusion if y < x. You will show in the next
assignment that the above condition implies continuity of F .
Now, we assume that f is continuous at some p ∈ (a, b). Let h ∈ R such that p + h ∈ (a, b).
Then,
F (p + h) − F (p) 1
Z p+h
= f (x)d x
h h p

1
Z p+h
= f (p) + f (x) − f (p)d x
h p

1
Z p+h
= f (p) + ( f (x) − f (p))d x
h p
= f (p) +G(h).

We wish to show that limh→0 G(h) = 0, in other words, given ε > 0, there is a δ > 0 such that
whenever 0 < |h| < δ, we have that |G(h)| < ε.
Let ε > 0. By the continuity of f at p, there exists a δ > 0 such that whenever x ∈ (a, b) and
|x − p| < δ, we have that | f (x) − f (p)| < ε. Thus, using the previous theorem
 ¯R ¯
¯ Z p+h
¯1
¯  1 ¯¯ p+h f (x) − f (p)d x ¯¯ , h > 0,
¯ h ¯p
¯ ( f (x) − f (p))d x ¯¯ = ¯
¯h
p − 1 ¯¯R p+h f (x) − f (p)d x ¯¯ , h < 0,
h p
 R
 1 p+h ¯ f (x) − f (p)¯ d x, h > 0,
¯ ¯
h p
≤ < ε.
− 1 R p ¯ f (x) − f (p)¯ d x, h < 0,
¯ ¯
h p+h


Remarks. (1) If f is not continuous at p, that does not necessarily imply that F is not differen-
tiable at p. For instance, in the next assignment, you will show that if

0, x ∈ [−1, 0) ∪ (0, 1],
f (x) =
1, x = 0,

then f is integrable,r and F (x) = 0 for all x ∈ [−1, 1]. However, F ′ (0) ̸= f (1). On the other hand,
consider 
0, x ∈ [−1, 0),
f (x) =
1, x ∈ [0, 1],

Then, 
and 0, x ∈ [−1, 0),
F (x) =
x, x ∈ [0, 1],
44

which is simply not differentiable at 0.


(2) The function F above is sometimes called an indefinite integral of f . If f is nonnegative,
then F is increasing, since
Z y
F (y) − F (x) = f (x)d x ≥ 0.
x
If f is increasing, then F inherits the property of “convexity". We will not discuss this in this
course, but you may read Section 2.18 from Apostol.
(3) One could have also defined
Z x
G(x) = f (t )d t .
b
Rb
Then, F (x) − G(x) = a f (t )d t , which is a constant. Thus, G has the same continuity and differ-
entiability properties of F !

Definition 5.13. Let f : (a, b) → R be a function. A differentiable function F : (a, b) → R is called a


primitive/anti-derivative of f on (a, b) if F ′ = f on (a, bW ).

Remark. If F and G are two primitives of f on (a, b), then (F − G)′ = 0. Thus, F = G + C for some
constant C .
x r +1
Example 28. sin(x) and cos(x) are primitives of cos(x) and − sin(x), respectively. r +1 is a primi-
r
tive of x on (0, ∞) for all r ∈ R \ {1}.
By the First FTOC, we also know that
Z x 1
F (x) = dt, x > 0,
1 t
1
is a primitive of f (x) = x
on (0, ∞).

Theorem 5.14 (The second Fundamental Theorem of Calculus). Let I be an open interval, ¯ and
f : I → R be a function that admits an anti-derivative F on I . Let [a, b] ⊆ I , and suppose f ¯ is
¯
[a,b]
Riemann integrable. Then,
Z b
f (x)d x = F (b) − F (a).
a

END OF LECTURE 31

Proof. We will show that


I − ε ≤ F (b) − F (a) ≤ I + ε
for all ε > 0.
45

Let ε > 0. By the integrability of f , there exist s ε ∈ S f and t ε ∈ T f such that


Z b
s ε (x)d x > I − ε,
a
Z b
t ε (x)d x < I + ε.
a

By taking common refinements, we may assume that s ε and t ε are step functions with respect
to the same partition P = {x 0 < · · · < x n }, i.e., there exist s 1 , ..., s n ∈ R and t 1 , ..., t n ∈ R such that
s ε (x) = s j and t ε (x) = t j for all x ∈ (x j −1 , x j ).
Now, by MVT, there exist c j ∈ (x j −1 , x j ) such that
n n n
F ′ (c j )(x j − x j −1 ) =
X X X
F (b) − F (a) = F (x j ) − F (x j −1 ) = f (c j )(x j − x j −1 ).
j =1 j =1 j =1

But, s j = s ε (c j ) ≤ f (c j ) ≤ t ε (c j ) = t j for all j = 1, ..., n. Thus,


Z b n n
I −ε ≤
X X
s ε (x)d x = s(c j )(x j − x j −1 ) ≤ f (c j )(x j − x j −1 )
a j =1 j =1
n Z b
t ε (x)d x < I + ε.
X
≤ t (c j )(x j − x j −1 ) =
j =1 a

Since ε > 0 was arbitrary, we are done. □

Remark. We can now write formulas such as


3b 2/3 3a 2/3
Z b Z b
1 −1/3
1/3
d x = x d x = − .
a x a 2 2
R
Remark (Leibniz’ notation). Leibniz used the notation f (x)d x to denote a general primitive of
f , e.g., the equation
Z
cos(x)d x = sin(x) +C

merely means that the derivative of sin(x) is cos(x). This symbol (without limits) has nothing to
R
do with integration a priori. You should simply read “ f (x)d x" as "the general primitive of f "
The two fundamental theorems tell us that:
R Rx
1. For a continuous f , f (x)d x = a f (t )d t +C .
Rb R ¯b
2. a f (x)d x = f (x)d x ¯
¯
a

The FTOC’s allow us to convert theorems about differentiabilty to theorems about integrabil-
¤′
ity. For instance, suppose F ′ = f . We know that F ◦ g (x) = F ′ (g (x))g ′ (x) = f (g (x))g ′ (x). Not
£
46

worrying about integrability for now, this should tell us that


Z b Z g (b) Z g (b)
′ ′
f (g (x))g (x)d x = F (g (b)) − F (g (a)) = F (x)d x = f (x)d x.
a g (a) g (a)

This is precisely integration by substitution!

Theorem 5.15 (Integration by substitution). Let I be an open interval, and g : I → R be a function


that admits a continuous derivative on I . Let J be an open interval containing g (I ), and f : J → R
be a continuous function. Then, for an [a, b] ⊆ I ,
Z b Z g (b)

f (g (x))g (x)d x = f (x)d x.
a g (a)

Proof. We will find a primitive of f (g (x))g ′ (x) on [a, b]. Let


Z x
F (x) = f (t )d t , x ∈ J .
g (a)

The above integral exist due the continuity of f of J . By the FFTOC, we have that F is differen-
tiable on J , and
F ′ (x) = f (x), x ∈ J.
Since g (I ) ⊂ J , we have that F ◦ g is well-defined and differentiable on I . By the chain rule,

(F ◦ g )′ (x) = F ′ (g (x))g ′ (x) = f (g (x))g ′ (x), x ∈ I.

for all x ∈ [a, b]. Thus, F ◦ g is an anti-derivative of ( f ◦ g )g ′ on I . By the SFTOC, applied twice,
Z b Z g (b)

f (g (x))g (x)d x = F (g (b)) − F (g (a)) = f (x)d x.
a g (a)

One can similarly obtain the formula for Integration by Parts as a consequence of the Funda-
mental Theorems of Calculus and the Product Rule for derivatives, but we will skip that.
END OF LECTURE 32

5.1. Logarithm and exponentiation. For any b > 0, we wish to define logb (x) as that real number
z such that

(5.1) b z = x.

However, in this course, we are yet to properly define what b z means if z is an irrational number.
Moreover, why should such a z exist? Why should it be unique. Let us extract the key features of
the desired log function from (5.1).
47

Say, z = " logb (x)" and w = " logb (y)", i.e., b z = x and b w = y. Then, if powers are to behave as
expected, b z+w = x y, i.e., logb (x) + logb (y) = z + w = " logb (x y)". Thus, we seek a function ℓ such
that
ℓ(x y) = ℓ(x) + ℓ(y)
for all x, y in its domain. Some observations based on this functional equation:
• If 0 ∈ dom(ℓ), then taking x = y = 0, we have that ℓ(0) = 2ℓ(0) ⇒ ℓ(0) = 0. Taking only
y = 0, we have that ℓ(0) = ℓ(x) for all x ∈ dom(ℓ). Thus, we only get the constant zero
function.
• If 1, −1 ∈ dom(ℓ), then taking x = y = 1 gives that

ℓ(1) = 0,

and taking x = y = −1 yields that ℓ(−1) = 0.


• Now, if x, −x, 1, −1 ∈ dom(ℓ), taking y = −1 yields that f (x) = f (−x). Thus, f is an even
function on its domain.
• Assuming ℓ is differentiable, taking y as a constant and differentiating in x, we obtain that

yℓ′ (y x) = ℓ′ (x).

Evaluating at x = 1 yields that


ℓ′ (1)
ℓ′ (y) = .
y
By FTOC, we would obtain that

ℓ(y) − ℓ(1) = ℓ′ (1) R y d t , y > 0,
1 t
ℓ(y) =
ℓ(y) − ℓ(−1) = ℓ′ (1) R −y d t , y < 0.
1 t

The choice ℓ (1) gives the trivial solution. For any non-zero choice of ℓ′ (1), ℓ(x)/ℓ′ (1) gives

yet another function that satisfies the functional equation (f.e.)!

Definition 5.16. Let x > 0. The natural logarithm of x is defined as


Z x
dt
ln(x) = .
1 t

Theorem 5.17. The function ln : R → has the following properties:


• ln(1) = 0,
• ln(x y) = ln(x) + ln(y), x, y > 0,
• ln is increasing and continuous on (0, ∞)
• ln(x) is differentiable on (0, ∞), and ln′ (x) = x1 , x > 0,
• (in Leibniz notation) dtt = ln |t | +C , t ̸= 0,
R
48
R
• (Leibniz) ln(x)d x = x ln(x) − x +C ,
• ln is one-to-one and onto R.

Proof. □

Definition 5.18. Let e denote the unique positive number such that ln(e) = 1. For any x ∈ R, let
exp(x) denote the unique positive number y such that ln(y) = x, i.e., the function exp(x) is the
inverse of ln(x).

The constant e was defined earlier using an infinite series. This turns out to be the same e!

Theorem 5.19. The function exp : R → R has the following properties:


• exp(0) = 1,
• exp(x + y) = exp(x) exp(y), x, y ∈ R,
• exp is increasing and continuous on R,
• exp is differentiable on R and exp′ (x) = exp(x), x ∈ R,
R
• (Leibniz) exp(x)d x = exp(x) +C ,
• exp is one-to-one and onto (0, ∞)
• exp(r ) = e r for every rational r .

Proof. □

Definition 5.20. Let a > 0 and x ∈ R. Define

a x = exp(x ln a).

END OF LECTURE 33

You might also like