DG
DG
Riemannian Geometry
F REDERICK T SZ -H O F ONG
Preface ix
vii
viii Contents
Bibliography 285
Preface
This lecture note is written for courses MATH 4033 (Calculus on Manifolds) and MATH
6250I (Riemannian Geometry) taught by the author in the Hong Kong University of
Science and Technology.
The main goal of these courses is to introduce advanced undergraduates and first-
year graduate students the basic concepts of differentiable manifolds, tensor calculus,
cohomology, and Riemannian geometry. It presents some of the most essential knowledge
on differential geometry that is necessary for further studies or research in geometric
analysis, general relativity, string theory, and related fields. Before reading the lecture
notes and taking these courses, students are advised to have a solid conceptual background
of linear algebra (MATH 2131) and multivariable calculus.
The course MATH 4033 covers Chapters 1 to 5 in this lecture note. These chapters
are about the analytic, algebraic, and topological aspects of differentiable manifolds.
Chapters 6 and 7 form a crush course on differential geometry of hypersurfaces in
Euclidean spaces. The main purpose of these two chapters is to give some motivations
on why various abstract concepts in Riemannian geometry are introduced in the way
they are. The remaining Chapters 8 to 11 form an introduction course to Riemannian
geometry.
Students are very welcome to point out typographical errors of any part of the notes,
and contribute diagrams to the Riemannian geometry chapters. The author would like
to thank the following students for their diligent readings of the earlier version of the
lecture notes and for pointing out many typographical errors: Chow Ka-Wing, Alex
Chan Yan-Long, Aaron Chow Tsz-Kiu, Jimmy Choy Ka-Hei, Toby Cheung Hin-Wa, Poon
Wai-Tung, Cheng Chun-Kit, Chu Shek-Kit, Wan Jingbo, Nicholas Chin Cheng-Hoong, Tang
Tianchen, and Luk Hoi-Ping.
Frederick Tsz-Ho Fong
January 18, 2021
HKUST, Clear Water Bay, Hong Kong
ix
Part 1
Differentiable Manifolds
Chapter 1
Regular Surfaces
Wolfgang Pauli
where 0 < θ < 2π and 0 < ϕ < π. This parametrization covers almost every part
of the sphere (except the north and south poles, and a half great circle connecting
them). In order to cover the whole sphere, we need more parametrizations, such as
G(θ, ϕ) = (sin ϕ cos θ, sin ϕ sin θ, cos ϕ) with domain −π < θ < π and 0 < ϕ < π.
Since the image of either F or G does not cover the whole sphere (although almost),
from now on we call them local parametrizations.
3
4 1. Regular Surfaces
To many students (myself included), the definition of regular surfaces looks obnox-
ious at the first glance. One way to make better sense of it is to look at some examples
and understand why each of the three conditions is needed in the definition.
The motivation behind condition (1) in the definition is that we are studying differen-
tial topology/geometry and so we want the parametrization to be differentiable as many
times as we like. Condition (2) rules out surfaces that have self-intersection such as the
Klein bottle (see Figure 1.2a). Finally, condition (3) guarantees the existence of a unique
tangent plane at every point on M (see Figure 1.2b for a non-example).
1.1. Local Parametrizations 5
(a) Klein Bottle has a self-intersection. (b) F (u, v) = (u3 , v 3 , uv) fails condition (3).
Exercise 1.1. Show that F (u, v) : (0, 2π) × (0, 1) → R3 defined by:
F (u, v) = (sin u, sin 2u, v)
satisfies conditions (1) and (3) in Definition 1.1, but not condition (2). [Hint: Try
to show F −1 is not continuous by finding a diverging sequence {(un , vn )} such that
{F (un , vn )} converges. See Figure 1.4 for reference.]
In Figure 1.3, one can observe that there are two families of curves on the surface.
These curves, often called coordinate curves, are obtained by varying one of the (u, v)-
variables while keeping the other constant. Precisely, they are the curves represented by
F (u, v0 ) and F (u0 , v) where u0 and v0 are fixed. As such, the partial derivatives ∂F
∂u (p)
∂F
and ∂v (p) give a pair of tangent vectors on the surface at point p. Therefore, their
cross product ∂F ∂F
∂u (p) × ∂v (p) is a normal vector to the surface at point p (see Figure
1.5). Here we have abused the notations for simplicity: ∂F ∂F
∂u (p) means ∂u evaluated at
(u, v) = F −1 (p). Similarly for ∂F
∂v (p).
Condition (3) requires that ∂F ∂F
∂u × ∂v is everywhere non-zero in the domain of F . An
equivalent statement is that the vectors ∂F ∂F
∂u (p), ∂v (p) are linearly independent for any
p ∈ F (U).
However, not all points on the sphere are covered by S2+ and S2− , since points on the
equator are not. In order to show that S2 is a regular surface, we need to write down
more smooth local parametrization(s) so that each point on the sphere can be covered by
at least one smooth local parametrization chart. One can construct four more smooth
local parametrizations (left, right, front and back) similar to F1 and F2 (see Figure 1.6).
It is left as an exercise for readers to write down the other four parametrizations. These
six parametrizations are all smooth and they cover the whole sphere. Therefore, it shows
the sphere is a regular surface.
Exercise 1.2. Write down the left, right, front and back parametrizations Fi ’s
(i = 3, 4, 5, 6) of the sphere as shown in Figure 1.6. Indicate clearly the domain and
range of each Fi .
8 1. Regular Surfaces
Example 1.5 (Sphere: revisited). We can in fact cover the sphere by just two smooth
local parametrizations described below. Define F+ (u, v) : R2 → S2 \{(0, 0, 1)} where:
u2 + v 2 − 1
2u 2v
F+ (u, v) = , ,
u2 + v 2 + 1 u2 + v 2 + 1 u2 + v 2 + 1
It is called the stereographic parametrization of the sphere (see Figure 1.7) , which assigns
each point (u, v, 0) on the xy-plane of R3 to a point where the line segment joining
(u, v, 0) and the north pole (0, 0, 1) intersects the sphere. Clearly F+ is a smooth function.
We leave it as exercise for readers to verify that F+ satisfies condition (3) and that
F+−1 : S2 \{(0, 0, 1)} → R2 is given by:
−1 x y
F+ (x, y, z) = , .
1−z 1−z
As z 6= 1 for every (x, y, z) in the domain of F+−1 , it is a continuous function. Therefore,
F+ is a smooth local parametrization. The inverse map F+−1 is commonly called the
stereographic projection of the sphere.
Note that the range of F+ does not include the point (0, 0, 1). In order to show
that the sphere is a regular surface, we need to cover it by another parametrization
F− : R2 → S2 \{(0, 0, −1)} which assigns each point (u, v, 0) on the xy-plane to a point
where the line segment joining (u, v, 0) and the south pole (0, 0, −1) intersects the sphere.
It is an exercise for readers to write down the explicit parametrization F− .
Exercise 1.3. Verify that F+ in Example 1.4 satisfies condition (3) in Definition 1.1,
and that the inverse map F+−1 : S2 \{(0, 0, 1)} → R2 is given as stated. [Hint: Write
down F+ (u, v) = (x, y, z) and solve (u, v) in terms of (x, y, z). Begin by finding
u2 + v 2 in terms of z.]
Furthermore, write down explicitly the map F− described in Example 1.4, and
find its inverse map F−−1 .
Exercise 1.4. Find smooth local parametrizations which together cover the whole
ellipsoid:
x2 y2 z2
2
+ 2 + 2 =1
a b c
where a, b and c are positive constants.
1.1. Local Parametrizations 9
Let’s also look at a non-example of smooth local parametrizations. Consider the map:
G(u, v) = (u3 , v 3 , 0), (u, v) ∈ R × R.
It is a smooth, injective map from R onto the xy-plane Π of R3 , i.e. G : R2 → Π.
2
Proof. The key idea of the proof is to use the Implicit Function Theorem. Given any
point p = (x0 , y0 , z0 ) ∈ g −1 (c), since ∇g(x0 , y0 , z0 ) 6= (0, 0, 0), at least one of the first
partials:
∂g ∂g ∂g
(p), (p), (p)
∂x ∂y ∂z
is non-zero. Without loss of generality, assume ∂g ∂z (p) 6= 0, then the Implicit Function
Theorem shows that locally around the point p, the level set g −1 (c) can be regarded as a
graph z = f (x, y) of some smooth function f of (x, y). To be precise, there exists an open
set O of g −1 (c) containing p such that there is a smooth function f (x, y) : U ⊂ R2 → R
from an open set U such that (x, y, f (x, y)) ∈ O ⊂ g −1 (c) for any (x, y) ∈ U. As such, the
smooth local parametrization F : U → O defined by:
F (u, v) = (u, v, f (u, v))
is a smooth local parametrization of g −1 (c).
∂g
In the case where ∂y (p) 6= 0, the above argument is similar as locally around p one
−1
can regard g (c) as a graph y = h(x, z) for some smooth function h. Similar in the case
∂g
∂x (p) 6= 0.
Since every point p can be covered by the image of a smooth local parametrization,
the level set g −1 (c) is a regular surface.
Example 1.7. The unit sphere x2 +y 2 +z 2 = 1 is a level surface g −1 (1) where g(x, y, z) :=
x2 + y 2 + z 2 . The gradient vector ∇g = (2x, 2y, 2z) is zero only when (x, y, z) = (0, 0, 0).
Since the origin is not on the unit sphere, we have ∇g(x0 , y0 , z0 ) 6= (0, 0, 0) for any
(x0 , y0 , z0 ) ∈ g −1 (1). Therefore, the unit sphere is a regular surface.
Similarly, one can also check that the surface x2 + y 2 = z 2 + 1 is a regular surface.
It is a level set h−1 (1) where h(x, y, z) = x2 + y 2 − z 2 . Since ∇h = (2x, 2y, −2z), the
origin is the only point p at which ∇h(p) = (0, 0, 0) and it is not on the level set h−1 (1).
Therefore, h−1 (1) is a regular surface.
Exercise 1.6. [dC76, P.66] Let f (x, y, z) = (x + y + z − 1)2 . For what values of c is
the set f −1 (c) a regular surface?
The proof of Theorem 1.6 makes use of the Implicit Function Theorem which is an
existence result. It shows a certain level set is a regular surface, but it fails to give an
explicit smooth local parametrization around each point.
There is one practical use of Theorem 1.6 though. Suppose we are given F (u, v)
which satisfies conditions (1) and (3) in Definition 1.1 and that F is continuous and
F −1 exists. In order to verify that it is a smooth local parametrization, we need to prove
continuity of F −1 , which is sometimes difficult. Here is one example:
F (u, v) = (sin u cos v, sin u sin v, cos u), 0 < u < π, 0 < v < 2π
is a smooth local parametrization of a unit sphere. It is clearly a smooth map from
(0, π) × (0, 2π) ⊂ R2 to R3 , and it is quite straight-forward to verify condition (3) in
Definition 1.1 and that F is one-to-one. However, it is rather difficult to write down an
explicit F −1 , let alone to show it is continuous.
The following result tells us that if the surface is given by a level set satisfying
conditions stated in Theorem 1.6, and F satisfies conditions (1) and (3), then F −1 is
automatically continuous. Precisely, we have the following:
Proposition 1.8. Assume all given conditions stated in Theorem 1.6. Furthermore, suppose
F (u, v) is a bijective map from an open set U ⊂ R2 to an open set O ⊂ M := g −1 (c) which
satisfies conditions (1) and (3) in Definition 1.1. Then, F satisfies condition (2) as well
and hence is a smooth local parametrization of g −1 (c).
Proof. Given any point p ∈ g −1 (c), we can assume without loss of generality that
∂g
∂z (p) 6= 0. Recall from Multivariable Calculus that ∇g(p) is a normal vector to the level
surface g −1 (c) at point p. Furthermore, if F (u, v) is a map satisfying conditions (1) and
−1
(3) of Definition 1.1, then ∂F ∂F
∂u (p) × ∂v (p) is also a normal vector to g (c) at p.
Now that the k̂-component of ∇g(p) is non-zero since ∂g
∂z (p) 6= 0, so the k̂-component
∂F ∂F
of the cross product ∂u (p) × ∂v (p) is also non-zero. If we express F (u, v) as:
F (u, v) = (x(u, v), y(u, v), z(u, v)),
∂F ∂F
then the k̂-component of ∂u (p) ×(p) is given by:
∂v
∂x ∂x
∂x ∂y ∂y ∂x
− (p) = ∂u
∂y
∂v
∂y (p).
∂u ∂v ∂u ∂v ∂u ∂v
Define π : R3 → R2 by π(x, y, z) = (x, y). The above shows that the composition π ◦ F
given by
(π ◦ F )(u, v) = (x(u, v), y(u, v))
has non-zero Jacobian determinant at p. By the Inverse Function Theorem, π ◦ F has a
smooth local inverse near p. In particular, (π ◦ F )−1 is continuous near p.
Finally, by the fact that (π ◦ F ) ◦ F −1 = π and that (π ◦ F )−1 exists and is continuous
locally around p, we can argue that F −1 = (π ◦ F )−1 ◦ π is also continuous near p. It
completes the proof.
12 1. Regular Surfaces
∂g
Exercise 1.8. Rewrite the proof of Proposition 1.8 by assuming ∂y (p) 6= 0 instead.
Example 1.9. We have already shown that the unit sphere x2 + y 2 + z 2 = 1 is a regular
surface using Theorem 1.6 by regarding it is the level set g −1 (1) where g(x, y, z) =
x2 + y 2 + z 2 . We also discussed that
F (u, v) = (sin u cos v, sin u sin v, cos u), 0 < u < π, 0 < v < 2π
is a possible smooth local parametrization. It is clearly smooth, and by direct computation,
one can show
∂F ∂F
× = sin u (sin u cos v, sin u sin v, cos u)
∂u ∂v
∂F ∂F
and so ∂u × ∂v = sin u 6= 0 for any (u, v) in the domain (0, π) × (0, 2π). We leave it as
an exercise for readers to verify that F is one-to-one (and so bijective when its codomain
is taken to be its image).
Condition (2) is not easy to verify because it is difficult to write down the inverse map
F −1 explicitly. However, thanks for Proposition 1.8, F is a smooth local parametrization
since it satisfies conditions (1) and (3), and it is one-to-one.
Exercise 1.9. Consider that the Mercator projection of the unit sphere:
cos v sin v sinh u
F (u, v) = , ,
cosh u cosh u cosh u
where sinh u := 21 (eu − e−u ) and cosh u := 21 (eu + e−u ).
(a) What are the domain and range of F ?
(b) Show that F is a smooth local parametrization.
Note that Fα−1 (W) and Fβ−1 (W) are open subsets of Uα and Uβ respectively. The
map Fβ−1 ◦ Fα describes a relation between two sets of coordinates (u1 , u2 ) and (v1 , v2 )
of M . In other words, one can regard Fβ−1 ◦ Fα as a change-of-coordinates, or transition
map and we can write:
Fβ−1 ◦ Fα (u1 , u2 ) = (v1 (u1 , u2 ), v2 (u1 , u2 )).
One goal of this section is to show that this transition map Fβ−1 ◦ Fα is smooth
provided that Fα and Fβ are two overlapping smooth local parametrizations. Before we
present the proof, let us look at some examples of transition maps.
14 1. Regular Surfaces
Example 1.10. The xy-plane Π in R3 is a regular surface which admits a global smooth
parametrization Fα (x, y) = (x, y, 0) : R2 → Π. Another way to locally parametrize Π is
by polar coordinates Fβ : (0, ∞) × (0, 2π) → Π
Fβ (r, θ) = (r cos θ, r sin θ, 0)
Readers should verify that they are smooth local parametrizations. The image of Fα
is the entire xy-plane Π, whereas the image of Fβ is the xy-plane with the origin and
positive x-axis removed. The transition map Fα−1 ◦ Fβ is given by:
Fα−1 ◦ Fβ : (0, ∞) × (0, 2π) → R2 \{(x, 0) : x ≥ 0}
(r, θ) 7→ (r cos θ, r sin θ)
To put it in a simpler form, we can say (x(r, θ), y(r, θ)) = (r cos θ, r sin θ).
Exercise 1.12. The unit cylinder Σ2 in R3 can be covered by two local parametriza-
tions:
F : (0, 2π) × R → Σ2 Fe : (−π, π) × R → Σ2
F (θ, z) := (cos θ, sin θ, z) Fe(θ,
e ze) := (cos θ,
e sin θ,
e ze)
Compute the transition maps F −1 ◦ Fe and Fe−1 ◦ F . State their maximum possible
domains. Are they smooth on their domains?
Exercise 1.13. The Möbius strip Σ2 in R3 can be covered by two local parametriza-
tions:
F : (−1, 1) × (0, 2π) → Σ2 Fe : (−1, 1) × (−π, π) → Σ2
θe
θ
3 + u cos 2 cos θ
3 + u
e cos 2 cos θe
F (u, θ) = 3 + u cos θ2 sin θ Fe(e
u, θ)
e = 3+u
e cos θ2 sin θe
e
u sin θ2 θe
u
e sin 2
Compute the transition maps, state their maximum possible domains and verify that
they are smooth.
The proposition below shows that the transition maps between any pair of smooth
local parametrizations are smooth:
Proof. It suffices to show Fβ−1 ◦ Fα is smooth as the other one Fα−1 ◦ Fβ can be shown
by symmetry. Furthermore, since differentiability is a local property, we may fix a point
p ∈ W ⊂ M and show that Fβ−1 ◦ Fα is smooth at the point Fα−1 (p).
1.3. Transition Maps 15
Exercise 1.14. Rewrite the proof of Proposition 1.11, mutatis mutandis, by assuming
∂(y, z)
det (p) 6= 0 instead.
∂(u1 , u2 )
16 1. Regular Surfaces
Now let M and N be two regular surfaces in R3 . Then, one can also talk about
mappings Φ : M → N between them. In this section, we will define the notion of smooth
maps between two surfaces.
Suppose F : UM → M and G : UN → N are two smooth local parametrizations of M
and N respectively. One can then consider the composition G−1 ◦ Φ ◦ F after shrinking
the domain. It is then a map between open subsets of R2 .
However, in order for this composition to be well-defined, we require the image
of Φ ◦ F to be contained in the image of G, which is not always guaranteed. Let
W := Φ(OM ) ∩ ON be the overlapping region on N of these two images. Then, provided
6 ∅, the composition G−1 ◦ Φ ◦ F becomes well-defined as a map on:
that W =
From now on, whenever we talk about this composition G−1 ◦ Φ ◦ F , we always implicitly
assume that W 6= ∅ and its domain is (Φ ◦ F )−1 (W).
Definition 1.17 (Maps of Class C k ). Let M and N be two regular surfaces in R3 , and
Φ : M → N be a map between them. We say Φ is C k at p ∈ M if for any smooth local
parametrization F : UM → M with p ∈ F (UM ), and G : UN → N with Φ(p) ∈ G(UN ),
the composition G−1 ◦ Φ ◦ F is C k at F −1 (p) as a map between subsets of R2 .
If Φ is C k at p for any p ∈ M , then we say that Φ is C k on M . Here k can be taken
to be ∞, and in such case we call Φ to be C ∞ (or smooth) on M .
= (−u1 , −u2 )
Clearly, the map (u1 , u2 ) 7→ (−u1 , −u2 ) is C ∞ . It shows the antipodal map Φ is C ∞ at
every point in F1 (B1 (0)). One can show in similar way using other local parametrizations
that Φ is C ∞ at points on S2 not covered by F1 .
Note that, for instance, the images of Φ ◦ F1 and F1 are disjoint, and so F1−1 ◦ Φ ◦ F1
is not well-defined. We don’t need to verify whether it is smooth.
Exercise 1.16. Let Φ be the antipodal map considered in Example 1.19, and F+
and F− be the two stereographic parametrizations of S2 defined in Example 1.5.
Compute the maps F+−1 ◦ Φ ◦ F+ , F−−1 ◦ Φ ◦ F+ , F+−1 ◦ Φ ◦ F− and F−−1 ◦ Φ ◦ F− .
State their domains, and verify that they are smooth on their domains.
Exercise 1.18. Given any pair of C ∞ functions f, g : R2 → R, show that the graphs
Γf and Γg are diffeomorphic.
Exercise 1.20. Show that the equation of the tangent plane p + Tp M of the graph
of a smooth function f (x, y) at p = (x0 , y0 , f (x0 , y0 )) is given by:
∂f ∂f
z = f (x0 , y0 ) + (x − x0 ) + (y − y0 )
∂x (x0 ,y0 ) ∂y (x0 ,y0 )
Exercise 1.21. [dC76, P.88] Consider the surface M given by z = xf (y/x), where
x 6= 0 and f is a smooth function. Show that the tangent planes p + Tp M must pass
through the origin (0, 0, 0).
∂(Φ◦F )
derivatives ∂ui :
∂Φ ∂(Φ ◦ F ) d
(Φ(p)) := = Φ ◦ F ((u1 , u2 ) + têi )
∂ui ∂ui (u1 ,u2 ) dt t=0
where (u1 , u2 ) is a point in UM such that F (u1 , u2 ) = p. The curve F ((u1 , u2 ) + têi ) is a
curve on M with parameter t along the ui -direction. The curve Φ ◦ F ((u1 , u2 ) + têi ) is
then the image of the ui -curve of M under the map Φ (see Figure 1.11). It is a curve on
∂Φ
N so ∂u i
which is a tangent vector to the surface N .
∂F
Then, one can regard Φ∗ as a map that takes the tangent vector ∂u i
in Tp M to another
n o
∂Φ ∂F
vector ∂uj in TΦ(p) N . Since ∂ui (p) is a basis of Tp M , one can then extend Φ∗ linearly
and define it as the tangent map of Φ. Precisely, we have:
22 1. Regular Surfaces
If the point p is clear from the context, (Φ∗ )p can be simply denoted by Φ∗ .
Remark 1.25. Some textbooks may use dΦp to denote the tangent map of Φ at p.
Example 1.26. Consider the unit sphere S2 locally parametrized by
F (θ, ϕ) = (sin ϕ cos θ, sin ϕ sin θ, cos ϕ)
and the rotation map:
Φ(x, y, z) = (x cos α − y sin α, x sin α + y cos α, z)
From Exercise 1.22, one should have figured out that:
∂Φ
= (− sin ϕ sin(θ + α), sin ϕ cos(θ + α), 0)
∂θ
∂Φ
= (cos ϕ cos(θ + α), cos ϕ sin(θ + α), − sin ϕ)
∂ϕ
∂F ∂F
Next we want to write them in terms of the basis , . However, we should
∂θ ∂ϕ
be careful about the base points of these vectors. Consider a point p ∈ S2 with local
∂Φ ∂Φ
coordinates (θ, ϕ), the vectors and computed above are based at the point Φ(p)
∂θ ∂ϕ
with
local coordinates (θ+ α, ϕ). Therefore, weshould express them in terms of the basis
∂F ∂F ∂F ∂F
(Φ(p)), (Φ(p)) , not (p), (p) !
∂θ ∂ϕ ∂θ ∂ϕ
At Φ(p), we have:
∂F ∂Φ
(Φ(p)) = (− sin ϕ sin(θ + α), sin ϕ cos(θ + α), 0) = (Φ(p))
∂θ ∂θ
∂F ∂Φ
(Φ(p)) = (cos ϕ cos(θ + α), cos ϕ sin(θ + α), − sin ϕ) = (Φ(p))
∂ϕ ∂ϕ
Therefore, the tangent map (Φ∗ )p acts on the basis vectors by:
∂F ∂F
(Φ∗ )p (p) = (Φ(p))
∂θ ∂θ
∂F ∂F
(Φ∗ )p (p) = (Φ(p))
∂ϕ ∂ϕ
In other words, the matrix representation [(Φ∗ )p ] with respect to the bases
∂F ∂F 2 ∂F ∂F
(p), (p) for Tp S (Φ(p)), (Φ(p)) for TΦ(p) S2
∂θ ∂ϕ ∂θ ∂ϕ
is the identity matrix. However, it is not perfectly correct to say (Φ∗ )p is an identity map,
since the domain and co-domain are different tangent planes.
1.5. Tangent Planes and Tangent Maps 23
Example 1.27. Let Φ : S2 → S2 be the rotation map as in Example 1.26. Consider again
the local parametrization:
F (θ, ϕ) = (sin ϕ cos θ, sin ϕ sin θ, cos ϕ).
By standard trigonometry, one can find out that Φ(F (θ, ϕ)) = F (θ + α, ϕ). Equivalently,
the map F −1 ◦ Φ ◦ F (in a suitable domain) is the map:
(θ, ϕ) 7→ (θ + α, ϕ).
As α is a constant, the Jacobian matrix of F −1n◦ Φ ◦ F ois the identity
n matrix,
o and
so the matrix [(Φ∗ )p ] with respect to the bases ∂F ,
∂θ ∂ϕ
∂F
and ∂F ∂F
,
∂θ ∂ϕ is the
p Φ(p)
identity matrix (which was also obtained by somewhat tedious computations in Example
1.26).
Abstract Manifolds
Shmuel Weinberger
25
26 2. Abstract Manifolds
Given two topological spaces (X, TX ) and (Y, TY ), one can talk about functions or
mapping between them. A map Φ : X → Y is said to be continuous with respect to TX
and TY if for any U ∈ TY , we have Φ−1 (U ) ∈ TX . This definition is a generalization of
continuous functions between Euclidean spaces equipped with the usual topologies. If
the map Φ : X → Y is one-to-one and onto, and both Φ and Φ−1 are continuous, then
we say Φ is a homeomorphism and the spaces (X, TX ) and (Y, TY ) are homeomorphic.
A topological space (X, T ) is said to be Hausdorff if for any pair of distinct points
p, q ∈ X, we have U1 , U2 ∈ T such that p ∈ U1 , q ∈ U2 and U1 ∩ U2 = ∅. In other words,
points of a Hausdorff space can be separated by open sets. It is intuitive that Rn with the
usual topology is a Hausdorff space. Any subset S ⊂ Rn with subspace topology is also a
Hausdorff space.
A topological space (X, TX ) is said to be second countable if there is a countable
sub-collection {Ui }∞ i=1 ⊂ T such that any set U ∈ T can be expressed as a union of some
of these Ui ’s. For instance, Rn with usual topology is second countable since by density
of rational numbers, any open set can be expressed as a countable union of open balls
with rational radii and centers.
This introduction to point-set topology is intended to be short. It may not make sense
to everybody, but it doesn’t hurt! Point-set topology is not the main dish of the course.
Many spaces we will look at are either Euclidean spaces, their subsets or sets derived
from Euclidean spaces. Most of them are Hausdorff and second countable. Readers
who want to learn more about point-set topology may consider taking MATH 4225. For
more thorough treatment on point-set topology, please consult [Mun00]. Meanwhile,
the take-home message of this introduction is that we don’t have to worry much about
point-set topology in this course!
2.1.2. Definitions and Examples. Now we are ready to learn what a manifold is.
We will first introduce topological manifolds, which are objects that locally look like
Euclidean space in certain continuous sense:
Example 2.4. Any regular surface is a topological manifold since its local parametriza-
tions are all homeomorphisms. Therefore, spheres, cylinders, torus, etc. are all topologi-
cal manifolds.
However, a double cone (see Figure 2.1) is not a topological manifold since the
vertex is a “bad” point. Any open set containing the vertex cannot be homeomorphic to
any open set in Euclidean space.
2.1. Smooth Manifolds 27
Figure 2.1. Double cone is not locally Euclidean near its vertex.
Remark 2.5. Note that around every p there may be more than one local parametriza-
tions. If Fα : Uα → Oα and Fβ : Uβ → Oβ are two local parametrizations around p, then
the composition:
are often called the transition maps between these local parametrizations. We need
to restrict their domains to smaller sets so as to guarantee the transition maps are
well-defined (c.f. Section 1.3).
Remark 2.7. Two local parametrizations Fα and Fβ with smooth transition maps Fα−1 ◦Fβ
and Fβ−1 ◦ Fα are said to be compatible.
Remark 2.8. We often use the superscript n, i.e. M n , to mean that the manifold M is
n-dimensional.
Example 2.10. Any topological manifold which can be covered by one global parametriza-
tion (i.e. image of F is all of M ) is a smooth manifold. Examples of which include Rn
which can be covered by one parametrization Id : Rn → Rn . The graph of Γf of any con-
tinuous function f : Rn → R is also a smooth manifold covered by one parametrization
F (x) = (x, f (x)) : Rn → Γf . Any regular curve γ(t) is a smooth manifold of dimension
1.
2.1. Smooth Manifolds 29
Example 2.11. All regular surfaces in R3 are smooth manifolds by Proposition 1.11
(which we showed their transition maps are smooth). Therefore, spheres, cylinders, tori,
etc. are all smooth manifolds.
Example 2.12 (Extended complex plane). Define M = C ∪ {∞}. One can show (omitted
here) that it is a Hausdroff, second countable topological space. Furthermore, one can
cover M by two local parametrizations:
F1 : R2 → C ⊂ M F2 : R2 → (C\{0}) ∪ {∞} ⊂ M
1
(x, y) 7→ x + yi (x, y) 7→
x + yi
The overlap part on M is given by C\{0}, corresponding to R2 \{(0, 0)} in R2 under
the parametrizations F1 and F2 . One can compute that the transition maps are given by:
−1 x y
F2 ◦ F1 (x, y) = ,− 2
x2 + y 2 x + y2
−1 x y
F1 ◦ F2 (x, y) = ,− 2
x2 + y 2 x + y2
Both are smooth maps on R2 \{(0, 0)}. Therefore, C ∪ {∞} is a smooth manifold.
Exercise 2.2. Discuss: According to Example 2.10, the graph of any continuous
function f : Rn → R is a smooth manifold as there is no transition map. However,
wouldn’t it imply the single cone:
n p o
(x, y, z) ∈ R3 : z = x2 + y 2
is a smooth manifold? It appears to have a “corner” point at the vertex, isn’t it?
2.1.3. Product and Quotient Manifolds. Given two smooth manifolds M m and
N , one can form an (m + n)-dimensional manifold M m × N n , which is defined by:
n
M m × N n := {(x, y) : x ∈ M m and y ∈ N n }.
Given a local parametrization F : UM → OM for M m , and a local parametrizaiton
G : UN → ON for N n , one can define a local parametrization:
F × G : UM × UN → OM × ON ⊂ M m × N n
(u, v) 7→ (F (u), G(v))
If {Fα } is a collection of local parametrizations of M m with smooth transition maps, and
{Gβ } is that of N n with smooth transition maps, then one can form a collection of local
parametrizations Fα × Gβ of the product M m × N n . It can be shown that these local
parametrizations of M m × N n also have smooth transition maps between open subsets
of Rm+n (see Exercise 2.3).
Exercise 2.3. Show that if Fα and Fαe are local parametrizations of M m with smooth
transition maps, and similarly for Gβ and Gβe for N n , then Fα × Gβ and Fαe × Gβe
have smooth transition maps.
30 2. Abstract Manifolds
The result from Exercise 2.3 showed that the product M m × N n of two smooth
manifolds M m and N n is a smooth manifold with dimension m + n. Inductively, the
product M1m1 × . . . Mkmk of k smooth manifolds M1m1 , . . . , Mkmk is a smooth manifold
with dimension m1 + . . . + mk .
Example 2.13. The cylinder x2 + y 2 = 1 in R3 can be regarded as R × S1 . The torus
can be regarded as S1 × S1 . They are both smooth manifolds. By taking products of
known smooth manifolds, one can generate a great deal of new smooth manifolds. The
n-dimensional cylinder can be easily seen to be a smooth manifold by regarding it as
R × Sn−1 . The n-dimensional torus S1 × . . . × S1 is also a smooth manifold.
| {z }
n times
Another common way to produce a new manifold from an old one is to take quotients.
Take R as an example. Let us define an equivalence relation ∼ by declaring that x ∼ y if
and only if x − y is an integer. For instance, we have 3 ∼ 5 while 4 6∼ 92 . Then, we can
talk about equivalence classes [x] which is the following set:
[x] := {y ∈ R : y ∼ x}.
For instance, we have 5 ∈ [2] as 5 ∼ 2. Likewise −3 ∈ [2] as −3 ∼ 2 too. The set [2] is the
set of all integers. Similarly, one can also argue [−1] = [0] = [1] = [2] = [3] = . . . are all
equal to the set of all integers.
On the contrary, 1 6∈ [0.2] as 1 6∼ 0.2. Yet −1.8, −0.8, 0.2, . . . are all in the set [0.2].
The set [0.2] is simply the set of all numbers in the form of 0.2 + N where N is any integer.
One can also see that [−1.8] = [−0.8] = [0.2] = [1.2] = . . ..
Under such notations, we see that [1] = [2] while [1] 6= [0.2]. The notion of equiva-
lence classes provides us with a way to “decree” what elements in the “mother” set (R in
this case) are regarded as equal. This is how topologists and geometers interpret gluing.
In this example, we can think of 1, 2, 3, etc. are glued together, and also −1.8, −0.8, 0.2,
etc. are glued together. Formally, we denote
R/ ∼ := {[x] : x ∈ R}
which is the set of all equivalence classes under the relation ∼. This new set R/ ∼ is
called a quotient set of R by the equivalence relation ∼. By sketching the set, we can see
R/ ∼ is topologically a circle S1 (see Figure 2.3):
Exercise 2.4. Describe the set R2 / ∼ where we declare (x1 , y1 ) ∼ (x2 , y2 ) if and
only if x1 − x2 ∈ Z and y1 − y2 ∈ Z.
Example 2.14 (Real Projective Space). The real projective space RPn is the quotient
set of Rn+1 \{0} under the equivalence relation: (x0 , x1 , . . . , xn ) ∼ (y0 , y1 , . . . , yn ) if and
only if there exists λ ∈ R\{0} such that (x0 , x1 , . . . , xn ) = (λy0 , λy1 , . . . , λyn ). Each
equivalence class is commonly denoted by:
[x0 : x1 : · · · : xn ]
For instance, we have [0 : 1 : −1] = [0 : −π : π]. Under this notation, we can write:
RPn := [x0 : x1 : · · · : xn ] : (x0 , x1 , . . . , xn ) ∈ Rn+1 \{0}
One can compute that the transition map F1−1 ◦ F0 is given by:
−1 1 x2 xn
F1 ◦ F0 (x1 , . . . , xn ) = , ,...,
x1 x1 x1
which is smooth on the domain F0−1 (O0 ∩ O1 ). The smoothness of transition maps
between any other pairs can be verified in a similar way.
Exercise 2.5. Express the transition map F3−1 ◦F1 of RP5 and verify that it is smooth
on its domain.
Example 2.15 (Complex Projective Space). The complex projective space CPn is an
important manifold in Complex Geometry (one of my research interests) and Algebraic
Geometry. It is defined similarly as RPn , with all R’s replaced by C’s. Precisely, we
declare for any two elements in (z0 , . . . , zn ), (w0 , . . . , wn ) ∈ Cn+1 \{(0, . . . , 0)}, we have
(z0 , . . . , zn ) ∼ (w0 , . . . , wn ) if and only if there exists λ ∈ C\{0} such that zi = λwi for
any i = 0, . . . n. Under this equivalence relation, the equivalence classes denoted by
[z0 : z1 : · · · : zn ] constitute the complex projective space:
CPn := {[z0 : z1 : · · · : zn ] : zi not all zero } .
It can be shown to be a smooth 2n-manifold in exactly the same way as in RPn .
Example 2.16. The Klein Bottle K (see Figure 1.2a) cannot be put inside R3 without
self-intersection, but it can be done in R4 . It is covered by two local parametrizations
given below:
These local parametrizations need to be compatible with each other in a sense that any
overlapping parametrizations Fα and Fβ must have smooth transition maps Fα−1 ◦ Fβ
and Fβ−1 ◦ Fα . Such a collection of local parametrizations A = {Fα , Uα , Oα }α is called a
smooth atlas of M .
Given a smooth atlas A of M , we can enlarge the atlas by including more local
parametrizations Fnew : Unew → Onew that are compatible to all local parametrizations in
A. The differential structure generated by an atlas A is a gigantic atlas that contains all
local parametrizations which are compatible with every local parametrizations in A (for
more formal definition, please read [Lee09, Section 1.3]).
Let’s take the plane R2 as an example. It can be parametrized by at least three
different ways:
• the identity map F1 := id : R2 → R2 .
2.1. Smooth Manifolds 33
Exercise 2.9. Show that any smooth manifold has uncountably many distinct
differential structures. [Hint: Let B(1) := {x ∈ Rn : |x| < 1}, consider maps
s
Ψs : B(1) → B(1) defined by Ψs (x) = |x| x where s > 0.]
34 2. Abstract Manifolds
Definition 2.17 (Functions and Maps of Class C k ). Let M m and N n be two smooth
manifolds of dimensions m and n respectively. Then:
A scalar-valued function f : M → R is said to be C k at p ∈ M if for any smooth
local parametrization F : U → M with p ∈ F (U), the composition f ◦ F is C k at the
point F −1 (p) ∈ U as a function from subset from Rm to R. Furthermore, if f : M → R
is C k at every p ∈ M , then we say f is C k on M .
A map Φ : M → N is said to be C k at p ∈ M if for any smooth local parametrization
F : UM → OM ⊂ M with p ∈ F (UM ), and G : UN → ON ⊂ N with Φ(p) ∈ G(UN ),
the composition G−1 ◦ Φ ◦ F is C k at F −1 (p) as a map between subsets of Rm and Rn .
Furthermore, if Φ : M → N is C k at every p ∈ M , then Φ is said to be C k on M .
When k is ∞, we can also say that the function or map is smooth.
Remark 2.18. By the definition of a smooth manifold (see condition (2) in Definition
2.6), transition maps are always smooth. Therefore, although we require f ◦ F and
G−1 ◦ Φ ◦ F to be smooth for any local parametrizations around p, it suffices to show
that they are smooth for at least one F covering p and at least one G covering Φ(p).
The domain of G−1 ◦ Φ ◦ F is (Φ ◦ F )−1 Φ ◦ F (R3 ) ∩ G(R2 ) , and the map is explicitly
given by:
G−1 ◦ Φ ◦ F (u1 , u2 , u3 )
−1 + k u2k
P
2u1 2u2 2u3
= G−1 ◦ Φ P 2, P 2, P 2, P 2
1 + k uk 1 + k uk 1 + k uk 1 + k uk
−1 + k u2k
P
−1 2u1 2u2 2u3
=G +i : +i
1 + k u2k 1 + k u2k 1 + k u2k 1 + k u2k
P P P P
" #
2u3 + i −1 + k u2k
P
= G−1 1 :
2u1 + 2iu2
2u1 u3 + u2 (−1 + k u2k ) −2u2 u3 + u1 (−1 + k u2k )
P P
= G−1 1 : + i
2(u21 + u22 ) 2(u21 + u22 )
2u1 u3 + u2 (−1 + k uk ) −2u2 u3 + u1 (−1 + k u2k )
P 2 P
= , .
2(u21 + u22 ) 2(u21 + u22 )
For any (u1 , u2 , u3 ) in the domain of G−1 ◦Φ◦F , which is (Φ◦F )−1 Φ ◦ F (R3 ) ∩ G(R2 ) ,
Exercise 2.11. Suppose Φ : Rn+1 \{0} → Rm+1 \{0} is a smooth map which satisfies
Φ(cx0 , cx1 , . . . , cxn ) = cd Φ(x0 , x1 , . . . , xn )
for any c ∈ R\{0} and (x0 , x1 , . . . , xn ) ∈ Rn+1 \{0}. Show that the induced map
e : RPn → RPm defined by:
Φ
Φe ([x0 : x1 : · · · : xn ]) = Φ(x0 , x1 , . . . , xn )
is well-defined and smooth. [Hint: To check Φ e is well-defined means to verify that
two equivalent inputs [x0 : x1 : · · · : xn ] = [y0 : y1 : · · · : yn ] will give the same
outputs Φ(x0 , x1 , . . . , xn ) and Φ(y0 , y1 , . . . , yn ).]
2 2
Exercise 2.12. Let M = {(w, z) ∈ C2 : |w| + |z| = 1}.
(a) Show that M is a 3-dimensional manifold.
(b) Define
2 2
Φ(w, z) := z w̄ + wz̄, i(wz̄ − z w̄), |z| − |w|
for any (w, z) ∈ M . Show that Φ(w, z) ∈ R3 and it lies on the unit sphere S2 ,
and then verify that Φ : M → S2 is a smooth map.
id−1 −1
U ◦ π ◦ F (x) = idU ◦ π(x, f (x))
= id−1
U (x)
= x.
Therefore, the composite id−1
◦ π ◦ F is simply the identity map on U, which is clearly
U
smooth.
π is one-to-one and onto with inverse map π −1 given by:
π −1 : U → Γf
x 7→ (x, f (x))
To show π −1 is smooth, we consider the composite F −1 ◦ π −1 ◦ idU :
F −1 ◦ π −1 ◦ idU (x) = F −1 ◦ π −1 (x)
= F −1 (x, f (x))
= x.
Therefore, the composite F −1 ◦ π −1 ◦ idU is also the identity map on U, which is again
smooth.
Example 2.23. Let M be the cylinder x2 + y 2 = 1 in R3 . We are going to show that M
is diffeomorphic to R2 \{(0, 0)} via the diffeomorphism:
Φ : M → R2 \{(0, 0)}
(x, y, z) 7→ ez (x, y)
We leave it for readers to verify that Φ is one-to-one and onto, and hence Φ−1 exists. To
show it is a diffeomorphism, we first parametrize M by two local coordinate charts:
F1 : (0, 2π) × R → M F2 : (−π, π) × R → M
F1 (θ, z) = (cos θ, sin θ, z) F2 (θ,
e ze) = (cos θ,
e sin θ,
e ze)
The target space R2 \{(0, 0)} is an open set of R2 , and hence can be globally parametrized
by id : R2 \{(0, 0)} → R2 \{(0, 0)}.
We need to show Φ ◦ Fi and Fi−1 ◦ Φ−1 are smooth for any i = 1, 2. As an example,
we verify one of them only:
Φ ◦ F1 (θ, z) = Φ(cos θ, sin θ, z)
= (ez cos θ, ez sin θ).
38 2. Abstract Manifolds
To show F1−1 ◦ Φ−1 = (Φ ◦ F1 )−1 is smooth, we use Inverse Function Theorem. The
Jacobian of Φ ◦ F1 is given by:
z
−e sin θ ez cos θ
D(Φ ◦ F1 ) = det z = −e2z 6= 0.
e cos θ ez sin θ
Therefore, Φ ◦ F1 has a C ∞ local inverse around every point in the domain. Since Φ ◦ F1
is one-to-one and onto, such a local inverse is a global inverse.
Similarly, one can show Φ ◦ F2 and F2−1 ◦ Φ−1 are smooth. All these show Φ and Φ−1
are smooth maps between M and R2 \{(0, 0)}, and hence are diffeomorphisms.
Exercise 2.13. Show that the open square (−1, 1) × (−1, 1) ⊂ R2 is diffeomorphic
to R2 . [Hint: consider the trig functions tan or tan−1 .]
∂
The partial derivative (p) can be thought as an operator:
∂uj
∂
(p) : C 1 (M, R) → R
∂uj
∂f
f 7→ (p).
∂uj
Here C 1 (M, R) denotes the set of all C 1 functions from M to R.
∂F ∂F
On regular surfaces (p) is a tangent vector at p. On an abstract manifold, (p)
∂uj ∂uj
cannot be defined since F may not be in an Euclidean space. Instead, the partial
∂ ∂F
differential operator (p) plays the role of (p), and we will call the operator
∂uj ∂uj
∂
(p) a tangent vector for an abstract manifold. It sounds strange to call a derivative
∂uj
operator a tangent vector. For beginners, you may try to get used to it by fantasizing the
∂
letter F whenever you see . “F” stands for “fantasize” (or in Hong Kong’s slang: try
∂uj
∂
to “FF” there were an F in )!
∂uj
Example 2.25. Let F (x, y, z) = (x, y, z) be the identity parametrization of R3 , and
G(ρ, θ, ϕ) = (ρ sin ϕ cos θ, ρ sin ϕ sin θ, ρ cos ϕ) be local parametrization of R3 by spherical
coordinates.
40 2. Abstract Manifolds
∂ ∂ ∂F ∂F
Figure 2.5. ∂u
and ∂v
are used in place of ∂u
and ∂v
on abstract manifolds.
∂ ∂ ∂
Then at any point p ∈ R3 , the vectors (p), (p), (p) are regarded as the
∂x ∂y ∂z
∂F ∂F ∂F
abstract form of the geometric vectors (p), (p), (p), which are respectively î, ĵ
∂x ∂y ∂z
and k̂ in standard notations.
∂ ∂ ∂
Also, the vectors (p), (p), (p) are regarded as the abstract form of the geo-
∂ρ ∂θ ∂ϕ
∂G ∂G ∂G
metric vectors (p), (p), (p), which are respectively the vectors at p tangent to
∂ρ ∂θ ∂ϕ
the ρ-, θ- and ϕ-directions on the sphere.
F : R2 → RP2
(x1 , x2 ) 7→ [1 : x1 : x2 ]
∂F ∂F
Such a manifold is not assumed to be in RN , so we can’t define , as geometric
∂x1 ∂x2
∂ ∂
vectors in RN . However, as a substitute, we will regard the operators , as abstract
∂x1 ∂x2
tangent vectors along the directions of x1 and x2 respectively.
∂ ∂ 0
In other words, can be expressed as a linear combination of s.
∂ui ∂vj
n n
∂ ∂
Therefore, span (p) ⊂ span (p) . Since both spans of vectors have
∂ui i=1 ∂vi i=1
equal dimension, their span must be equal. This shows Tp M is independent of choice of
local parametrizations. However, it is important to note that each individual basis vector
∂
(p) does depend on local parametrizations.
∂ui
Example 2.28. Consider again the real projective plane:
RP2 = {[x0 : x1 : x2 ] : at least one xi 6= 0}.
Consider the two local parametrizations:
F : R2 → RP2 G : R2 → RP2
F (x1 , x2 ) = [1 : x1 : x2 ] G(y0 , y2 ) = [y0 : 1 : y2 ]
Then, (y0 , y2 ) can be regarded as a function of (x1 , x2 ) via the transition map G−1 ◦ F ,
which is explicitly given by:
(y0 , y2 ) = G−1 ◦ F (x1 , x2 ) = G−1 ([1 : x1 : x2 ])
−1 −1 −1 1 x2
= G ([x1 : 1 : x1 x2 ]) = , .
x1 x1
∂ ∂ ∂ ∂
Using the chain rule, we can then express , in terms of , :
∂x1 ∂x2 ∂y0 ∂y2
∂ ∂y0 ∂ ∂y2 ∂
= +
∂x1 ∂x1 ∂y0 ∂x1 ∂y2
1 ∂ x2 ∂
=− 2 − 2
x1 ∂y0 x1 ∂y2
∂ ∂
= −y02 − y0 y2 .
∂y0 ∂y2
∂
We leave as an exercise.
∂x2
∂ ∂ ∂
Exercise 2.16. Express as a linear combination , in Example 2.28.
∂x2 ∂y0 ∂y2
Leave the final answer in terms of y0 and y2 only.
Exercise 2.18. Given two smooth manifolds M m and N n , and a point (p, q) ∈
M × N , show that the tangent plane T(p,q) (M × N ) is isomorphic to Tp M ⊕ Tq N .
Recall that V ⊕ W is the direct sum of two vector spaces V and W , defined as:
V ⊕ W = {(v, w) : v ∈ V and w ∈ W }.
2.3. Tangent Spaces and Tangent Maps 43
2.3.3. Tangent Maps. Given a smooth map Φ between two regular surfaces in
R3 , we discussed in Section 1.5.2 on how to define its partial derivatives using local
parametrizations. To recap, suppose Φ : M → N and F (u1 , u2 ) : UM → OM ⊂ M and
G(v1 , v2 ) : UN → ON ⊂ N are local parametrizations of M and N respectively. Via Φ,
the local coordinates (v1 , v2 ) of N can be regarded as functions of (u1 , u2 ), i.e.
(v1 , v2 ) = G−1 ◦ Φ ◦ F (u1 , u2 ).
∂Φ
Then, according to (1.1), the partial derivative of the map Φ is given by:
∂ui
∂Φ ∂(Φ ◦ F ) ∂(G ◦ (G−1 ◦ Φ ◦ F ))
(Φ(p)) := =
∂ui ∂ui F −1 (p) ∂ui F −1 (p)
2
X ∂vj ∂G
= (Φ(p))
j=1
∂ui F −1 (p) ∂vj
∂Φ
Note that the partial derivative defined in (2.2) depends on the local parametriza-
∂ui
tion F . However, one can show that it does not depend on the choice of the local
parametrization G in the target space.
Suppose G(w
e 1 , . . . , wn ) is another local parametrization around Φ(p). Then by the
chain rule:
n n n
!
X ∂wj ∂ X ∂wj X ∂vk ∂
=
j=1
∂ui ∂wj j=1
∂ui ∂wj ∂vk
k=1
n
X ∂wj ∂vk ∂
=
∂ui ∂wj ∂vk
j,k=1
n
X ∂vk ∂
=
∂ui ∂vk
k=1
44 2. Abstract Manifolds
∂Φ
Therefore, the way to define in (2.2) is independent of choice of local parametrization
∂ui
G for the target manifold N .
Example 2.30. Consider the map Φ : RP1 × RP2 → RP5 defined by:
Φ([x0 : x1 ], [y0 : y1 : y2 ]) = [x0 y0 : x0 y1 : x0 y2 : x1 y0 : x1 y1 : x1 y2 ].
Under the standard local parametrizations F (u) = [1 : u] for RP1 , G(v1 , v2 ) = [1 : v1 : v2 ]
for RP2 , and H(w1 , . . . , w5 ) = [1 : w1 : · · · : w5 ] for RP5 , the local expression of Φ is
given by:
H −1 ◦ Φ ◦ (F × G)(u, v1 , v2 )
= H −1 ◦ Φ([1 : u], [1 : v1 : v2 ])
= H −1 ([1 : v1 : v2 : u : uv1 : uv2 ])
= (v1 , v2 , u, uv1 , uv2 ).
Via the map Φ, we can regard (w1 , w2 , w3 , w4 , w5 ) = (v1 , v2 , u, uv1 , uv2 ), and the partial
derivatives of Φ are given by:
∂Φ ∂w1 ∂ ∂w5 ∂ ∂ ∂ ∂
= + ... + = + v1 + v2
∂u ∂u ∂w1 ∂u ∂w5 ∂w3 ∂w4 ∂w5
∂Φ ∂w1 ∂ ∂w5 ∂ ∂ ∂
= + ... + = +u
∂v1 ∂v1 ∂w1 ∂v1 ∂w5 ∂w1 ∂w4
∂Φ ∂w1 ∂ ∂w5 ∂ ∂ ∂
= + ... + = +u
∂v2 ∂v2 ∂w1 ∂v2 ∂w5 ∂w2 ∂w5
Definition 2.31 (Tangent Maps). Under the same assumption stated in Definition 2.29,
the tangent map of Φ at p ∈ M denoted by (Φ∗ )p is defined as:
(Φ∗ )p : Tp M → TΦ(p) N
n
! n
X ∂ X ∂Φ
(Φ∗ )p ai (p) = ai (p)
i=1
∂u i i=1
∂ui
If the point p is clear from the context, (Φ∗ )p can be simply denoted by Φ∗ .
For brevity, we will from now on say “(u1 , . . . , um ) are local coordinates of M around
p” instead of saying in a clumsy way that “F : U → M is a local parametrization of M
around p and that (u1 , . . . , um ) are coordinates on U”.
Given a local coordinates (u1 , . . . , um ) around p, and local coordinates (v1 , . . . , vn )
around Φ(p), then from (2.2), the matrix representation of (Φ∗ )p with respect to bases
m n j=1,...,n
∂ ∂ ∂vj
(p) and (Φ(p)) is given by where i stands for the
∂ui i=1 ∂vj j=1 ∂ui i=1,...,m
column, and j stands for the row. The matrix is nothing but the Jacobian matrix:
∂(v1 , . . . , vn )
= D(G−1 ◦ Φ ◦ F ) F −1 (p) .
(2.3) [(Φ∗ )p ] =
∂(u1 , . . . , um ) F −1 (p)
2.3. Tangent Spaces and Tangent Maps 45
Example 2.32. Consider again the map Φ : RP1 × RP2 → RP5 in Example 2.30. Under
the local parametrizations considered in that example, we then have (for instance):
∂ ∂Φ ∂ ∂ ∂
Φ∗ = = + v1 + v2 .
∂u ∂u ∂w3 ∂w4 ∂w5
Using the results computed in Example 2.30, the matrix representation of Φ∗ is given by:
0 1 0
0 0 1
1 0 0
[Φ∗ ] =
v1 u 0
v2 0 u
Hence, Φ∗ is injective. Remark: To be rigorous, we have only shown (Φ∗ )p is injective at
any p covered by the local coordinate charts we picked. The matrix [Φ∗ ] using other local
coordinate charts can be computed in a similar way (left as an exercise).
Exercise 2.19. Consider the map Φ : RP1 × RP2 → RP5 defined as in Example 2.30.
This time, we use the local parametrizations
F (u) = [u : 1]
G(v0 , v2 ) = [v0 : 1 : v2 ]
H(w0 , w1 , w3 , w4 , w5 ) = [w0 : w1 : 1 : w3 : w4 : w5 ]
for RP1 , RP2 and RP5 respectively. Compute matrix representation of Φ∗ using these
local parametrizations.
Exercise 2.20. Note that in Definition 2.31 we defined Φ∗ using local coordinates.
Show that Φ∗ is independent of local coordinates. Precisely, show that if:
X ∂ X ∂
ai = bi
i
∂u i i
∂w i
where {ui } and {wi } are two local coordinates of M , then we have:
! !
X ∂Φ X ∂Φ X ∂ X ∂
ai = bi , which implies Φ∗ ai = Φ∗ bi .
i
∂ui i
∂wi i
∂ui i
∂wi
Exercise 2.21. The identity map idM of a smooth manifolds M takes any point
p ∈ M to itself, i.e. idM (p) = p. Show that its tangent map (idM )∗ at p is the identity
map on the tangent space Tp M .
Exercise 2.22. Consider two smooth manifolds M m and N n , and their product
M m × N n . Find the tangent maps (πM )∗ and (πN )∗ of projection maps:
πM : M × N → M
(x, y) 7→ x
πN : M × N → N
(x, y) 7→ y
46 2. Abstract Manifolds
To find the tangent map (Ψ ◦ Φ)∗ , we need to figure out how it acts on the basis
∂
vectors , and recall that it is defined (see (2.2)) as follows:
∂ui
k
∂ ∂(Ψ ◦ Φ) X ∂wj ∂
(Ψ ◦ Φ)∗ = = .
∂ui ∂ui j=1
∂ui ∂wj
Next, we use the (standard) chain rule for maps between Euclidean spaces:
k k Xn
X ∂wj ∂ X ∂wj ∂vl ∂
= .
j=1
∂ui ∂wj j=1
∂vl ∂ui ∂wj
l=1
Therefore, we get:
k Xn
∂ X ∂wj ∂vl ∂
(Ψ ◦ Φ)∗ = .
∂ui j=1
∂vl ∂ui ∂wj
l=1
∂
Next, we verify that Ψ∗ ◦ Φ∗ will give the same output:
∂ui
2.4. Inverse Function Theorem 47
n
∂ ∂Φ X ∂vl ∂
Φ∗ = =
∂ui ∂ui ∂ui ∂vl
l=1
n
! n
∂ X ∂vl ∂ X ∂vl ∂
Ψ∗ ◦ Φ∗ = Ψ∗ = Ψ∗
∂ui ∂ui ∂vl ∂ui ∂vl
l=1 l=1
n n k
X ∂vl ∂Ψ X ∂vl X ∂wj ∂
= =
∂ui ∂vl ∂ui j=1 ∂vl ∂wj
l=1 l=1
k Xn
X ∂wj ∂vl ∂
= .
j=1
∂vl ∂ui ∂wj
l=1
Therefore, we have:
∂ ∂
(Ψ ◦ Φ)∗ = Ψ∗ ◦ Φ∗
∂ui ∂ui
for any i, and hence (Ψ ◦ Φ)∗ = Ψ∗ ◦ Φ∗ .
Proof. Given that Φ is a diffeomorphism, the inverse map Φ−1 : N → M exists. Since
Φ−1 ◦ Φ = idM , using the chain rule and Exercise 2.21, we get:
Exercise 2.23. Given two diffeomorphic smooth manifolds M and N , what can you
say about dim M and dim N ?
Proof. The proof to be presented uses the Inverse Function Theorem for Euclidean spaces
and then extends it to smooth manifolds. For the proof of the Euclidean case, readers
may consult the lecture notes of MATH 3033/3043.
Let F be a local parametrization of M near p, and G be a local parametrization
of N near Φ(p). Given that (Φ∗ )p is invertible, by (2.3) we know that the following
Jacobian matrix D(G−1 ◦ Φ ◦ F ) is invertible at F −1 (p). By Inverse Function Theorem
for Euclidean spaces, there exist an open set UM ⊂ Rdim M containing F −1 (p), and an
open set UN ⊂ Rdim N containing G−1 (Φ(p)) such that:
G−1 ◦ Φ ◦ F UM
: UM → UN
is a diffeomorphism, i.e. the inverse F −1 ◦ Φ−1 ◦ G exists when restricted to UN and is
smooth.
Denote OM = F (UM ) and ON = G(UN ). By the definition of smooth maps, this
shows Φ|OM and Φ−1 O are smooth. Hence Φ|OM is a local diffeomorphism near p.
N
Figure 2.7. a helicoid is not globally diffeomorphic to R2 \{0}, but is locally diffeomor-
phic to R2 \{0}.
It is clear that Φ is not injective: for instance, Φ(cos 2π, sin 2π, 2π) = Φ(cos 0, sin 0, 0).
However, we can show that (Φ∗ )p is injective at each point p ∈ Σ.
The set R2 \{0} is open in R2 . The matrix [Φ∗ ] is the Jacobian matrix of Φ ◦ F :
Φ ◦ F (r, θ) = Φ(r cos θ, r sin θ, θ)
= (r cos θ, r sin θ)
cos θ −r sin θ
[Φ∗ ] = D(Φ ◦ F ) = .
sin θ r cos θ
As det[Φ∗ ] = r 6= 0, the linear map [Φ∗ ] is invertible. By Inverse Function Theorem, Φ is
a local diffeomorphism.
Exercise 2.25. Show that Sn and RPn are locally diffeomorphic via the map:
Φ(x0 , . . . , xn ) = [x0 : · · · : xn ].
2.5. Immersions and Submersions 51
Remark 2.39. As a linear map T : V → W between any two finite dimensional vector
spaces cannot be injective if dim V > dim W , an immersion Φ : M → N can only exist
when dim M ≤ dim N .
Example 2.40. The map Φ : R → S1 defined by:
Φ(t) = (cos t, sin t)
( )
∂
is an immersion. The tangent space of R at any point t0 is simply span . The
∂t t=t0
tangent map (Φ∗ )t0 is given by:
∂ ∂Φ
(Φ∗ )t0 = = (− sin t0 , cos t0 ) 6= 0.
∂t ∂t t=t0
52 2. Abstract Manifolds
Therefore, the “matrix” of Φ∗ is a one-by-one matrix with a non-zero entry. Clearly, there
is no free column and so Φ∗ is injective at every t0 ∈ R. This shows Φ is an immersion.
This example tells us that an immersion Φ is not necessary injective.
Example 2.41. Let M be a regular surface in R , then the inclusion map ι : M → R3 ,
2 3 2
defined as ι(p) = p ∈ R3 , is a smooth map, since for any local parametrization F (u1 , u2 )
of M 2 , we have ι ◦ F = F , which is smooth by definition (see p.4). We now show that ι
is an immersion:
∂F ∂(ι ◦ F ) ∂F
(ι∗ )p = = .
∂ui ∂ui F −1 (p) ∂ui F −1 (p)
Let F (u1 , u2 ) = (x1 (u1 , u2 ), x2 (u1 , u2 ), x3 (u1 , u2 )), then
3
∂F X ∂xj
= êi
∂ui j=1
∂ui
where {êi } is the standard basis of R3 . Therefore, the matrix of ι∗ is given by:
∂x1 ∂x1
∂u1
∂x2
∂u2
∂x2
[ι∗ ] = ∂u1 ∂u2 .
∂x3 ∂x3
∂u1 ∂u2
Proof. The proof uses the Inverse Function Theorem. By translation, we may assume that
F (0) = p and G(0) = Φ(p). Given that (Φ∗ )p is injective, there are n linearly independent
rows in the matrix [(Φ∗ )p ]. WLOG we may assume that the first m rows of [(Φ∗ )p ] are
linearly independent. As such, the matrix can be decomposed into the form:
A
[(Φp )∗ ] =
∗
where A is an invertible m × m matrix, and ∗ denotes any k × m matrix.
Now define ψ : Rm+k → Rm+k as:
(*)
ψ(u1 , . . . , um , um+1 , . . . , um+k ) = G−1 ◦ Φ ◦ F (u1 , . . . , um ) + (0, . . . , 0, um+1 , . . . , um+k ).
We claim that this is the map ψ that we want. First note that ψ(0) = G−1 ◦ Φ(p) = 0
by our earlier assumption. Next we show that ψ has a smooth inverse near 0. The
Jacobian matrix of this map at 0 is given by:
A 0
[(Dψ)0 ] = .
∗ Ik
As rows of A are linearly independent, it is easy to see then all rows of [(Dψ)0 ] are
linearly independent, and hence [(Dψ)0 ] is invertible. By Inverse Function Theorem, ψ is
locally invertible near 0, i.e. there exists an open set UeN ⊂ Rm+k containing 0 such that
the restricted map:
ψ|UeN : UeN → ψ(UeN ) ⊂ UN .
has a smooth inverse.
Finally, we verify that this is the map ψ that we want. We compute:
(G ◦ ψ)−1 ◦ Φ ◦ F (u1 , . . . , um ) = ψ −1 (G−1 ◦ Φ ◦ F )(u1 , . . . , um ) .
Exercise 2.31. Show that if M and N are two smooth manifolds of equal dimension,
then the following are equivalent:
(i) Φ : M → N is a local diffeomorphism.
(ii) Φ : M → N is an immersion.
(iii) Φ : M → N is a submersion.
Exercise 2.33. Show that the map Φ : Rn+1 \{0} → RPn defined as:
Φ(x0 , . . . , xn ) = [x0 : · · · : xn ]
is a submersion.
One nice property of a submersion Φ : M m → N n that locally around every p ∈ M ,
one can find special local parametrizations F of M near p, and G of N near Φ(p) such
that G−1 ◦ Φ ◦ F is a projection map. We will see later that this result will show any
level-set of Φ, if non-empty, must be a smooth manifold. Let’s state this result in a precise
way:
Proof. The proof uses again the Inverse Function Theorem. First by translation we may
assume that F (0) = p and G(0) = Φ(p). Given that (Φ∗ )p is surjective, there are n
linearly independent columns in the matrix [(Φ∗ )p ]. WLOG assume that they are the first
n columns, then [(Φ∗ )p ] is of the form:
[(Φ∗ )p ] = D(G−1 ◦ Φ ◦ F )0 = A ∗
56 2. Abstract Manifolds
Let φ−1 (u1 , . . . , un+k ) = (v1 , . . . , vn+k ), then φ(v1 , . . . , vn+k ) = (u1 , . . . , un+k ). From
(*), we get:
(G−1 ◦ Φ ◦ F (v1 , . . . , vn+k ), vn+1 , . . . , vn+k ) = φ(v1 , . . . , vn+k )
= (u1 , . . . , un , un+1 , . . . , un+k )
−1
which implies G ◦ Φ ◦ F (v1 , . . . , vn+k ) = (u1 , . . . , un ). Combine with previous result,
we get:
(G−1 ◦ Φ ◦ F ) φ−1 (u1 , . . . , un+k )
2.6. Submanifolds
In this section we talk about submanifolds. A subspace W of a vector space V is a subset
of V and is itself a vector space. A subgroup H of a group G is a subset of G and is
itself a group. It seems that a smooth submanifold N of a smooth manifold M might be
defined as a subset of M and is itself a smooth manifold. However, it is just one side of
the full story – we need more than that because we hope that the local coordinates of a
submanifold is in some sense compatible with the local coordinates of the manifold M .
Exercise 2.34. Complete the exercise stated in Example 2.50 that if F1 and F2 are
compatible parametrizations of M around p, then the induced parametrizations Fe1
and Fe2 of ΓΦ are also compatible.
58 2. Abstract Manifolds
Exercise 2.37. Show that any non-empty open subset N of a smooth manifold M
is a submanifold of M with dim N = dim M .
We require a submanifold to have the inclusion map being smooth because we
want to rule out some pathological cases. Consider the graph of an absolute function,
i.e. Γ = {(x, |x|) : x ∈ R}, and R2 . The graph Γ can be parametrized by a single
parametrization F : R → Γ defined by:
F (t) = (t, |t|).
Then, since Γ equipped with this single parametrization, it is considered as a smooth
manifold (although quite difficult to accept) since there is essentially no transition map.
However, we (fortunately) can show that Γ is not a submanifold of R2 (with usual
differential structure, parametrized by the identity map). It is because the inclusion map
is not smooth:
id−1
R2 ◦ ι ◦ F (t) = (t, |t|)
Therefore,
{G(u1 , . . . , un , 0, . . . , 0) : (u1 , . . . , un , 0, . . . , 0) ∈ U}
= {F (u1 , . . . , un ) : (u1 , . . . , un ) ∈ UN }
= ON = N ∩ OM
It completes our proof.
Proof. Using Theorem 2.48 (Submersion Theorem), given any point p ∈ Φ−1 (q) ⊂ M ,
there exist a local parametrization F : UM → OM of M near p, and a local parametriza-
tion G of N near Φ(p) = q, such that:
G−1 ◦ Φ ◦ F (u1 , . . . , un , un+1 , . . . , um ) = (u1 , . . . , un )
and that F (0) = p, G(0) = q.
We first show that Φ−1 (q) is a smooth manifold. Note that we have:
Φ (F (0, . . . , 0, un+1 , . . . , um )) = G(0, . . . , 0) = q.
Therefore, F (0, . . . , 0, un+1 , . . . , um ) ∈ Φ−1 (q). Hence, Φ−1 (q) can be locally parametrized
by Fe(un+1 , . . . , um ) := F (0, . . . , 0, un+1 , . . . , um ). One can also verify that compatible
F ’s gives compatible Fe’s. This shows Φ−1 (q) is a smooth manifold of dimension m − n.
To show it is a submanifold of M , we need to compute the tangent map ι∗ . First
consider the composition:
F −1 ◦ ι ◦ Fe(un+1 , . . . , um ) = F −1 (F (0, . . . , 0, un+1 , . . . , um )) = (0, . . . , 0, un+1 , . . . , um ).
The matrix [ι∗ ] with respect to local parametrizations Fe of Φ−1 (q), and F of M is given
by the Jacobian:
0
[ι∗ ] = [D(F −1 ◦ ι ◦ Fe)] =
I
which shows ι∗ is injective. Therefore, Φ−1 (q) is a submanifold of M .
Using Proposition 2.52, one can produce a lot of examples of manifolds which are
level-sets of smooth functions.
Example 2.53. In R4 , the set Σ := {x3 + y 3 + z 3 + w3 = 1} is a smooth 3-manifold. It
can be shown by consider Φ : R4 → R defined by:
Φ(x, y, z, w) = x3 + y 3 + z 3 + w3 .
Then, Σ = Φ−1 (1). To show it is a manifold, we show Φ is a submersion at every p ∈ Σ.
By direct computation, we get:
[Φ∗ ] = [3x2 3y 2 3z 2 3w2 ]
Since [Φ∗ ] = 0 only when (x, y, z, w) = (0, 0, 0, 0) which is not contained in Σ, we have
shown (Φ∗ )p is injective for any p ∈ Σ. By Proposition 2.52, we have proved Σ = Φ−1 (1)
is a smooth manifold of dimension 4 − 1 = 3.
Example 2.54. The set Mn×n (R) of all n × n real matrices can be regarded as R2n
equipped with the usual differential structure. Consider these subsets of Mn×n (R):
60 2. Abstract Manifolds
At A0 ∈ Φ−1 (I), we have AT0 A0 = I and so for any symmetric matrix B, we have:
∂
(Φ∗ )A0 = (AT0 Eij )T + AT0 Eij
∂xij
n n
1 X ∂ = 1
X
[A0 B]ij (AT0 Eij )T + AT0 Eij
(Φ∗ )A0 [A0 B]ij
2 i,j=1 ∂xij 2 i,j=1
T
n n
1 T X 1 X
= A0 [A0 B]ij Eij + AT0 [A0 B]ij Eij
2 i,j=1
2 i,j=1
1 T 1
= (A A0 B)T + (AT0 A0 B)
2 0 2
1 1
= B T + B = B.
2 2
Therefore, (Φ∗ )A0 is surjective. This shows Φ∗ is a submersion at every point A0 ∈
Φ−1 (I). This shows Sym(n, R) = Φ−1 (I) is a submanifold of Mn×n (R) of dimension
dim Mn×n (R) − dim Sym(n, R), which is n2 − n(n+1)
2 = n(n−1)
2 .
Exercise 2.39. Show that the subset Σ of R3 defined by the two equations below is
a 1-dimensional manifold:
x3 + y 3 + z 3 = 1
x+y+z =0
Exercise 2.41. Consider the map Φ : S3 \{(0, 0)} → CP1 defined by:
Φ(x1 , x2 , x3 , x4 ) := [x1 + ix2 : x3 + ix4 ].
−1
Show that Φ ([1 : 0]) is a smooth manifold of (real) dimension 1, and show that
Φ−1 ([1 : 0]) is diffeomorphic to a circle.
Chapter 3
Michio Kaku
63
64 3. Tensors and Differential Forms
Exercise 3.1. Given that V is a finite-dimensional real vector space, show that:
(a) V ∗ is a vector space
(b) dim V ∗ = dim V
(c) If B = {ei }ni=1 is a basis for V , then B ∗ := {ei }ni=1 is a basis for V ∗ .
Given T ∈ V ∗ and that T (ei ) = ai , verify that:
Xn
T = ai ei .
i=1
Occasionally (just for aesthetic purpose), dui p can be denoted as dui p . Moreover,
whenever p is clear from the context (or not significant), we may simply write dui and
∂
∂ui .
Note that both Bp and Bp∗ depend on the choice of local coordinates. Suppose
(v1 , . . . , vn ) is another local coordinates around p, then by chain rule we have:
n
∂ X ∂uk ∂
=
∂vj ∂vj ∂uk
k=1
n
∂ X ∂vk ∂
= .
∂uj ∂uj ∂vk
k=1
3.1. Cotangent Spaces 65
Exercise 3.3. Suppose F (u1 , . . . , un ) and G(v1 , . . . , vn ) are two local parametriza-
tions of a smooth manifold M . Let ω : M → T M be a smooth differential 1-form
such that on the overlap of local coordinates we have:
X X
ω= aj duj = bi dv i .
j i
Definition 3.4 (Tangent and Cotangent Bundles). Let M be a smooth manifold. The
tangent bundle, denoted by T M , is defined to be:
[
TM = ({p} × Tp M ) .
p∈M
p∈M
Proposition 3.5. Let M n be a smooth manifold. Suppose F and G are two overlapping
smooth local parametrizations of M , then their induced local parametrizations Fe and G
e
defined as in (3.2) on the tangent bundle T M are compatible, and also that F and G∗
e ∗ e
defined as in (3.3) on the cotangent bundle T ∗ M are also compatible.
Corollary 3.6. The tangent bundle T M and the cotangent bundle T ∗ M of a smooth
manifold M are both smooth manifolds of dimension 2 dim M .
3.2. Tangent and Cotangent Bundles 67
Remark 3.8. In the above definition, we used V (p) to be denote the element (p, Vp ) in
T M , and Vp to denote the vector in Tp M . We will distinguish between them for a short
while. After getting used to the notations, we will abuse the notations and use Vp and
V (p) interchangeably.
Remark 3.9. Note that a vector field can also be defined locally on an open set O ⊂ M .
In such case we say V is a C k on O if the map V : O → T M is C k .
The functions V i : F (U) ⊂ M → R are all locally defined and are commonly called the
components of V with respect to local coordinates (u1 , . . . , un ).
Let Fe(u1 , . . . , un ; V 1 , . . . , V n ) be the induced local parametrization of T M defined
as in (3.2). Then, one can verify that:
n
!
−1 −1
X
i ∂
F ◦ V ◦ F (u1 , . . . , un ) = F
e e F (u1 , . . . , un ), V (F (u1 , . . . , un ))
i=1
∂ui F (u1 ,...,un )
1 n
= u1 , . . . , un ; V (F (u1 , . . . , un )), . . . , V (F (u1 , . . . , un )) .
Therefore, Fe−1 ◦ V ◦ F (u1 , . . . , un ) is smooth if and only if the components V i ’s are all
smooth. Similarly for class C k . In short, a vector field V is smooth if and only if the
components V i in every its local expression:
n
!
X
i ∂
V (p) = p, V (p) (p)
i=1
∂ui
3.2.3. Differential 1-Forms. Differential 1-forms are the dual counterpart of vector
fields. It is essentially an assignment of a cotangent vector to each point on M . Precisely:
Remark 3.11. At this moment we use ω(p) to denote an element in {p} × Tp∗ M , and
ωp to denote an element in Tp∗ M . We will abuse the notations later on and use them
interchangeably, since such a distinction is unnecessary for many practical purposes.
where ωi : F (U) ⊂ M → R are locally defined functions and are commonly called the
components of ω with respect to local coordinates (u1 , . . . , un ). Similarly to vector fields,
one can show that ω is a C ∞ differential 1-form if and only if all ωi ’s are smooth under
any local coordinates in the atlas of M (see Exercise 3.6).
Therefore, one can think of Φ∗ is a map which takes a cotangent vector ωΦ(p) ∈ TΦ(p)
∗
N
to a cotangent vector (Φ∗ ω)p on Tp∗ M . As it is in the opposite direction to Φ : M → N ,
we call Φ∗ the pull-back whereas Φ∗ is called the push-forward.
Remark 3.15. In many situations, the points p and Φ(p) are clear from the context.
Therefore, we often omit the subscripts p and Φ(p) when dealing with pull-backs and
push-forwards.
Therefore, Φ∗ ω = dθ.
Example 3.17. Let M := R2 \{(0, 0)} (equipped with polar (r, θ)-coordinates) and
N = R2 (with (x, y)-coordinates), and define:
Φ:M →N
Φ(r, θ) := (r cos θ, r sin θ)
Hence, we have:
∗ ∂ ∂
(Φ dx) = dx Φ∗
∂r ∂r
∂ ∂
= dx (cos θ) + (sin θ)
∂x ∂y
= cos θ
∂ ∂
(Φ∗ dx) = dx Φ∗
∂θ ∂θ
∂ ∂
= dx −y +x
∂x ∂y
= −y = −r sin θ
We conclude:
Φ∗ dx = cos θ dr − r sin θ dθ.
Exercise 3.9. Denote (x1 , x2 ) the coordinates for R2 and (y1 , y2 , y3 ) the coordinates
for R3 . Define the map Φ : R2 → R3 by:
Φ(x1 , x2 ) = (x1 x2 , x2 x3 , x3 x1 ).
Compute Φ (dy ), Φ (dy ) and Φ∗ (dy 3 ).
∗ 1 ∗ 2
well-defined to denote:
(DX Y )p := Dγ 0 (t) Y, (DX f )p := Dγ 0 (t) f at t = 0
By the existence and uniqueness theorems of ODE, such a curve γ(t) exists uniquely
provided that the vector field X is C 1 .
Furthermore, it can also be checked that if γ1 , γ2 : (−ε, ε) → Rn are two curves
with γ1 (0) = γ2 (0) = p and with the same velocity vectors γ10 (0) = γ20 (0) at p, then it
is necessarily that Dγ10 f = Dγ20 f and Dγ10 Y = Dγ20 Y at p. Therefore, just the existence
theorem of ODE is sufficient to argue that DX Y and DX f are well-defined.
Exercise 3.11. Prove the above claim that Dγ10 f = Dγ20 f and Dγ10 Y = Dγ20 Y at p.
Remark 3.18. Consult any standard textbook about theory of ODEs for a proof of
existence and uniqueness of the curve γ(t) given any vector field X. Most standard
textbook uses contraction mapping to prove existence, and Gronwall’s inequality to prove
uniqueness.
Now let M be a smooth manifold, and X be a smooth vector field on M . Then, one
can also extend the existence and uniqueness theorem of ODE to manifolds to prove that
for any point p ∈ M , there exists a smooth curve γ(t) : (−ε, ε) → M on M with γ(0) = p
such that:
d
γ(t) = X(γ(t)).
dt
d ∂
Recall that dt γ(t) is defined as γ∗ ∂t . This curve γ is called the integral curve of X
passing through p. This extension can be justified by applying standard ODE theorems
on the local coordinate chart covering p. Then one solves for the integral curve within
this chart until the curve approaches the boundary of the chart (say at point q). Since
the boundary of one chart must be the interior of another local coordinate chart, one can
then continue solving for the integral curve starting from q.
2.5 2.5
2.0 2.0
1.5 1.5
1.0 1.0
0.5 0.5
0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
Now one can still talk about integral curves γ(t) given a vector field X on a manifold,
so one can define Dγ 0 (t) f and DX f in the same way as in Rn (as it makes perfect sense
to talk about f (γ(t + δ)) − f (γ(t)). However, it is not straight-forward how to generalize
the definitions of Dγ 0 (t) Y and DX Y where Y is a vector field on a manifold. The vectors
Y (γ(t + δ)) and Y (γ(t)) are at different based points, so one cannot make sense of
Y (γ(t + δ)) − Y (γ(t)).
72 3. Tensors and Differential Forms
One notion of differentiating a vector field by another one is called the Lie derivatives.
The key idea is to push-forward tangent vectors in a natural way so that they become
vectors at the same based point. Then, it makes sense to consider subtraction of vectors
and also derivatives.
To begin with, we denote the integral curves using a map. First fix a vector field X
on a manifold M . Then, given any point p ∈ M , as discussed before, one can find an
integral curve γ(t) so that γ(0) = p and γ 0 (t) = X(γ(t)). We denote this curve γ(t) by
Φt (p), indicating that it depends on p. Now, for any fixed t, we can view Φt : M → M as
a map. One nice way to interpret this map is to regard Φt (p) as the point on M reached
by flowing p along the vector field X for t unit time. As such, this map Φt is often called
the flow map.
There are many meaningful purposes of this interpretation of integral curves. Stan-
dard theory of ODE shows Φt is smooth as long as X is a smooth vector field. Moreover,
given s, t ∈ R and p ∈ M , we can regard Φs (Φt (p)) as the point obtained by flowing p
along X first for t unit time, then for s unit time. Naturally, one would expect that the
point obtained is exactly Φs+t (p). It is indeed true provided that X is independent of t.
Proposition 3.19. Given any smooth vector field X on a smooth manifold M , and denote
its flow map by Φt : M → M . Then, given any t, s ∈ R and p ∈ M , we have:
(3.5) Φt (Φs (p)) = Φt+s (p), or equivalently Φt ◦ Φs = Φt+s .
Consequently, for each fixed t ∈ R, the flow map Φt is a diffeomorphism with inverse Φ−t .
Proof. The proof is a direct consequence of the uniqueness theorem of ODE. Consider s
as fixed and t as the variable, then Φt (Φs (p)) and Φt+s (p) can be regarded as curves on
M . When t = 0, both curves pass through the point Φs (p). It remains to show that both
curves satisfy the same ODE, then uniqueness theorem of ODE guarantee that the two
curves must be the same. We leave the detail as an exercise for readers.
Exercise 3.13. Complete the detail of the above proof that the curves {Φt (Φs (p))}t∈R
and {Φt+s (p)}t∈R both satisfy the same ODE.
Now we are ready to introduce Lie derivatives of vector fields. Given two vector
fields X and Y , we want to develop a notion of differentiating Y along X, i.e. the rate of
change of Y when moving along integral curves of X. Denote the flow map of X by Φt .
Fix a point p ∈ M , we want to compare YΦt (p) with Yp . However, they are at different
base points, so we push-forward YΦt (p) so that it becomes a vector based at p. To do so,
the natural way is to push it forward by the map Φ−t as it maps tangent vectors at Φt (p)
to tangent vectors at Φ−t (Φt (p)) = p.
Definition 3.20 (Lie Derivatives of Vector Fields). Let X and Y be smooth vector fields
on a manifold M . We define the Lie derivative of X along Y by:
d
(LX Y )p := (Φ−t )∗ YΦt (p)
dt t=0
where Φt denotes the flow map of X.
3.2. Tangent and Cotangent Bundles 73
It sounds like a very technical definition that is very difficult to compute! Fortunately,
we will prove that LX Y is simply the commutator [X, Y ], to be defined below. First recall
that a vector field on a manifold is a differential operator acting on scalar functions f .
After differentiating f by a vector field Y , we get another scalar function Y (f ). Then,
we can differentiate Y (f ) by another vector field X and obtaining X(Y (f )) (which for
simplicity we denote it by XY f . The commutator, or the Lie brackets, measure the
difference between XY f and Y Xf :
Definition 3.21 (Lie Brackets). Given two vector fields X and Y on a manifold M , we
define the Lie brackets [X, Y ] to be the vector field such that for any smooth function
f : M → R, we have:
[X, Y ]f := XY f − Y Xf.
Remark 3.22. Suppose under local coordinates (u1 , . . . , un ), the vector fields X and Y
can be written as:
n n
X ∂ X ∂
X= Xi Y = Yi ,
i=1
∂ui i=1
∂ui
then [X, Y ] has following the local expression:
n
∂Y j ∂X j
X ∂
(3.6) [X, Y ] = Xi −Yi
i,j=1
∂ui ∂ui ∂u j
Exercise 3.14. Verify (3.6), i.e. show that for any smooth function f : M → R, we
have:
n j j
i ∂Y i ∂X ∂f
X
XY f − Y Xf = X −Y
i,j=1
∂u i ∂ui ∂uj
It appears that [X, Y ] is more like an algebraic operation whereas Lie derivative LX Y
is a differential operation. Amazingly, they are indeed equal!
Proposition 3.23. Let X and Y be smooth vector fields on a manifold M . Then, we have:
LX Y = [X, Y ].
Proof. Denote Φt to be the flow map of the vector field X. Fix a point p ∈ M and let
F (u1 , . . . , un ) : U → M be a local parametrization covering p. In order to compute LX Y
at p, we may assume that t is sufficiently small so that Φt (p) is also covered by F . Denote
that coordinate representation of Φt by:
F −1 ◦ Φt ◦ F (u1 , . . . , un ) = (vt1 (u1 , . . . , un ), . . . , vtn (u1 , . . . , un )).
74 3. Tensors and Differential Forms
In local coordinates, the flow map Φt is then related to X under the relation:
n n
X ∂vti ∂ X ∂
i
= X i (Φt (p)) i .
i=1
∂t ∂u Φt (p) i=1 ∂u Φt (p)
| {z } | {z }
∂Φt XΦt (p)
∂t p
n
∂vtj ∂Z k ∂X j
X
= δjk =⇒ Zij (p) = δij and i
δjk + δik = 0.
∂uk t=0 | {z } ∂t ∂uk t=0
k=1
from definition
It implies that:
∂Zij ∂X j
=− .
∂t t=0 ∂ui
Then, we can compute that:
n
∂ ∂ ∂ X ∂
(3.10) (Φt )−1
∗ (Φt (p)) = Zik (p)
∂t t=0 ∂ui ∂t t=0 k=1 ∂uk
n
X ∂X k ∂
=− (p).
∂ui ∂uk
k=1
Now we know the geometric meaning of two commuting vector fields X and Y , i.e.
[X, Y ] = 0. According to Proposition 3.23, it is equivalent to saying LX Y = 0 at any
p ∈ M . Then by the definition of Lie derivatives, we can conclude that:
(Φ−t )∗ (YΦt (p) ) = (Φ−0 )∗ (YΦ0 (p) ) = Yp for any t.
In other words, we have YΦt (p) = (Φt )∗ (Yp ) for any t, meaning that pushing Y at p
forward by the flow map Φt of X will yield the vector field Y at the point Φt (p). This
result can further extends to show the flow maps of X and Y commute:
Exercise 3.17. Let X and Y be two vector fields on M such that [X, Y ] = 0. Denote
Φt and Ψt be the flow maps of X and Y respectively, show that for any s, t ∈ R, we
have:
Φs ◦ Ψt = Ψt ◦ Φs .
Sketch a diagram to illustrate its geometric meaning.
Lie derivatives on 1-forms can be defined similarly as on vector fields, except that we
uses pull-backs instead of push-forwards this time.
Definition 3.24 (Lie Derivatives of Differential 1-Forms). Let X and a smooth vector
field and α be a smooth 1-form on a manifold M . Denote the flow map of X by Φt ,
then we define the Lie derivative of α at p ∈ M along X by:
d
(LX α)p := (Φt )∗ αΦt (p) .
dt t=0
76 3. Tensors and Differential Forms
One can compute similarly as in Proposition 3.23 that the Lie derivative of a 1-form
Xn
α= αi dui can be locally expressed as:
i=1
n
∂X i
X ∂αj
(3.11) LX α = Xi + αi duj
i,j=1
∂ui ∂uj
3.3.1. Tensor Products in Vector Spaces. Given two vector spaces V and W ,
their dual spaces V ∗ and W ∗ are vector spaces of all linear functionals T : V → R and
S : W → R respectively. Pick two linear functionals T ∈ V ∗ and S ∈ W ∗ , their tensor
product T ⊗ S is a map from V × W to R defined by:
T ⊗S :V ×W →R
(T ⊗ S)(X, Y ) := T (X) S(Y )
It is easy to verify that T ⊗ S is bilinear, meaning that it is linear at each slot:
(T ⊗ S) (a1 X1 + a2 X2 , b1 Y1 + b2 Y2 )
= a1 b1 (T ⊗ S)(X1 , Y1 ) + a2 b1 (T ⊗ S)(X2 , Y1 )
+ a1 b2 (T ⊗ S)(X1 , Y2 ) + a2 b2 (T ⊗ S)(X1 , Y2 )
Exercise 3.21. Show that the tensor product is bilinear in a sense that:
T ⊗ (α1 S1 + α2 S2 ) = α1 T ⊗ S1 + α2 T ⊗ S2
and similar for the T slot.
78 3. Tensors and Differential Forms
Let’s take the dual basis as an example to showcase the use of tensor products.
Consider a vector space V with a basis {ei }ni=1 . Let {ei }ni=1 be its dual basis for V ∗ . Then,
one can check that:
n
X
In other words, the sum of tensor products Aij ei ⊗ ej is the inner product on V
i,j=1
represented by the matrix [Akl ] with respect to the basis {ei }ni=1 of V . For example, when
n
X Xn
Akl = δkl , then Aij ei ⊗ ej = ei ⊗ ei . It is the usual dot product on V .
i,j=1 i=1
Exercise 3.22. Show that {ei ⊗ej }ni,j=1 is a basis for V ∗ ⊗V ∗ . What is the dimension
of V ∗ ⊗ V ∗ ?
To describe linear or multilinear map between two vector spaces V and W (where
W is not necessarily the one-dimensional space R), one can also use tensor products.
Given a linear functional f ∈ V ∗ and a vector w ∈ W , we can form a tensor f ⊗ w, which
is regarded as a linear map f ⊗ w : V → W defined by:
(f ⊗ w)(v) := f (v)w.
Let {ei } be a basis for V , and {Fj } be a basis for W . Any linear map T : V → W can
be expressed in terms of these bases. Suppose:
X j
T (ei ) = Ai Fj .
j
Then, we claim that T can be expressed using the following tensor notations:
X j
T = Ai ei ⊗ Fj
i,j
3.3. Tensor Products 79
Let’s verify this. Note that a linear map is determined by its action on the basis {ei } for
V . It suffices to show:
X j
Ai ei ⊗ Fj (ek ) = T (ek ).
i,j
as desired.
Generally, if T1 , . . . , Tk ∈ V ∗ and X ∈ V , then
T1 ⊗ · · · ⊗ Tk ⊗ X
is regarded to be a k-linear map from V × . . . × V to V , defined by:
T1 ⊗ · · · ⊗ Tk ⊗ X : V × . . . × V → V
| {z }
k
(T1 ⊗ · · · ⊗ Tk ⊗ X)(Y1 , . . . , Yk ) := T1 (Y1 ) · · · Tk (Yk ) X
Example 3.25. One can write the cross-product in R3 using tensor notations. Think of
the cross product as a bilinear map ω : R3 × R3 → R3 that takes two input vectors u and
v, and outputs the vector u × v. Let {e1 , e2 , e3 } be the standard basis in R3 (i.e. {î, ĵ, k̂}).
Then one can write:
ω = e1 ⊗ e2 ⊗ e3 − e2 ⊗ e1 ⊗ e3
+ e2 ⊗ e3 ⊗ e1 − e3 ⊗ e2 ⊗ e1
+ e3 ⊗ e1 ⊗ e2 − e1 ⊗ e3 ⊗ e2
One can check that, for instance, ω(e1 , e2 ) = e3 , which is exactly e1 × e2 = e3 .
∂ ∂
g , = gij .
∂ui ∂uj
n
X
l ∂
Rm(X, Y, Z) = Rijk dui (X) duj (Y ) duk (Z) .
∂ul
i,j,k,l=1
n
∂ ∂ ∂ X
l ∂
Rm , , = Rijk .
∂ui ∂uj ∂uk ∂ul
l=1
We call g a (2, 0)-tensor (meaning that it maps two vectors to a scalar), and Rm a
(3, 1)-tensor (meaning that it maps three vectors to one vector). In general, we can also
define (k, 0)-tensor ω on M which has the general form:
n
X
ωp = ωi1 i2 ···ik (p) dui1 p
⊗ · · · ⊗ duik p
i1 ,...,ik =1
∂
Here ωi1 i2 ···ik ’s are scalar functions. This tensor maps the tangent vectors ∂ui1 , . . . , ∂u∂i
k
to the scalar ωi1 i2 ...ik at the corresponding point.
Like the Rm-tensor, we can also generally define (k, 1)-tensor Ω on M which has the
general form:
n
X ∂
Ωp = Ωji1 i2 ···ik (p) dui1 p
⊗ · · · ⊗ duik p
⊗ (p)
i1 ,...,ik ,j=1
∂uj
where Ωji1 i2 ...ik ’s are scalar functions. This tensor maps the tangent vectors ∂
∂ui1 , . . . , ∂u∂i
k
where ωi1 i2 ...ik ’s are smooth scalar functions locally defined on F (U).
A smooth (k, 1)-tensor Ω on M is a k-linear map Ωp : Tp M × . . . × Tp M → Tp M at
| {z }
k
each p ∈ M such that under any local parametrization F (u1 , . . . , un ) : U → M , it can
be written in the form:
n
X ∂
Ωp = Ωji1 i2 ···ik (p) dui1 p ⊗ · · · ⊗ duik p ⊗ (p)
i ,...,i ,j=1
∂uj
1 k
where Ωji1 i2 ...ik ’s are smooth scalar functions locally defined on F (U).
Remark 3.27. Since Tp M is finite dimensional, from Linear Algebra we know (Tp M )∗∗
∂
is isomorphic to Tp M . Therefore, a tangent vector ∂u i
(p) can be regarded as a linear
∗
functional on cotangent vectors in Tp M , meaning that:
∂
duj p = δij .
∂ui p
Under this interpretation, one can also view a(k, 1)-tensor Ω as a(k + 1)-linear map Ωp :
Tp M × . . . × Tp M ×Tp∗ M → R, which maps ∂
∂ui1 , . . . , ∂u∂i , duj to Ωji1 i2 ...ik . However,
| {z } k
k
we will not view a (k, 1)-tensor this way in this course.
Generally, we can also talk about (k, s)-tensors,which is a (k + s)-linear map
Ωp :
Tp M × . . . × Tp M × Tp∗ M × . . . × Tp∗ M → R taking ∂
∂ui1 , . . . , ∂u∂i , duj1 , . . . , dujs to a
| {z } | {z } k
k s
scalar. However, we seldom deal with these tensors in this course.
Exercise 3.24. Let M be a smooth manifold with local coordinates (u1 , u2 ). Consider
the tensor products:
∂
T1 = du1 ⊗ du2 and T2 = du1 ⊗ .
∂u2
Which of the following is well-defined?
∂
(a) T1
∂u1
∂
(b) T2
∂u1
∂ ∂
(c) T1 ,
∂u1 ∂u2
∂ ∂
(d) T2 ,
∂u1 ∂u2
82 3. Tensors and Differential Forms
Exercise 3.25. let M be a smooth manifold with local coordinates (u1 , u2 ). The
linear map T : Tp M → Tp M satisfies:
∂ ∂ ∂
T = +
∂u1 ∂u1 ∂u2
∂ ∂ ∂
T = − .
∂u2 ∂u1 ∂u2
Express T using tensor products.
Example 3.28. Consider the extended complex plane M := C ∪ {∞} defined in Example
2.12. We cover M by two local parametrizations:
F1 : R2 → C ⊂ M F2 : R2 → (C\{0}) ∪ {∞} ⊂ M
1
(x, y) 7→ x + yi (u, v) 7→
u + vi
The transition maps on the overlap are given by:
x y
(u, v) = F2−1 ◦ F1 (x, y) = , −
x2 + y 2 x2 + y 2
u v
(x, y) = F1−1 ◦ F2 (u, v) = , −
u2 + v 2 u2 + v 2
Consider the (2, 0)-tensor ω defined using local coordinates (x, y) by:
2
+y 2 )
ω = e−(x dx ⊗ dy.
Using the chain rule, we can express dx and dy in terms of du and dv:
(u2 + v 2 ) du − u(2u du + 2v dv)
u
dx = d 2 2
=
u +v (u2 + v 2 )2
v 2 − u2 2uv
= 2 du − 2 dv
(u + v 2 )2 (u + v 2 )2
(u2 + v 2 )dv − v(2u du + 2v dv)
v
dy = −d = −
u2 + v 2 (u2 + v 2 )2
2 2
2uv v −u
=− 2 du + 2 dv
(u + v 2 )2 (u + v 2 )2
Therefore, we get:
2uv(u2 − v 2 ) (u2 − v 2 )2
dx ⊗ dy = du ⊗ du + du ⊗ dv
(u2 + v 2 )4 (u2 + v 2 )4
4u2 v 2 2uv(u2 − v 2 )
+ 2 dv ⊗ du + dv ⊗ dv
(u + v 2 )4 (u2 + v 2 )4
2
+y 2 )
Recall that ω = e−(x dx ⊗ dy, and in terms of (u, v), we have:
2 1
+y 2 ) − u2 +v
e−(x =e 2
.
3.3. Tensor Products 83
Exercise 3.26. Consider the extended complex plane C ∪ {∞} as in Example 3.28,
and the (1, 1)-tensor of the form:
2 2 ∂
Ω = e−(x +y ) dx ⊗ .
∂y
Express Ω in terms of (u, v).
written using the ui ’s coordinates, one can convert it to vα ’s coordinates by the chain
rule:
!
X X X ∂ui X ∂uj
g= gij dui ⊗ duj = gij dv α ⊗ dv β
i,j i,j α
∂v α ∂v β
β
X X ∂ui ∂uj α
= gij dv ⊗ dv β
i,j
∂vα ∂vβ
α,β
Exercise 3.27. Given that ui ’s and vα ’s are overlapping local coordinates of a smooth
manifold M . Using these coordinates, one can express the following (3, 1)-tensor in
two ways:
X
l ∂ X
eη dv α ⊗ dv β ⊗ dv γ ⊗ ∂
Rm = Rijk dui ⊗ duj ⊗ duk ⊗ = R αβγ
∂ul ∂vη
i,j,k,l α,β,γ,η
l η
Express Rijk in terms of Rαβγ ’s.
Exercise 3.28. Given that ui ’s and vα ’s are overlapping local coordinates of a smooth
manifold M . Suppose g and h are two (2, 0)-tensors expressed in terms of local
coordinates as:
X X
g= gij dui ⊗ duj = geαβ dv α ⊗ dv β
i,j α,β
X X
i j
h= hij du ⊗ du = hαβ dv α ⊗ dv β .
e
i,j α,β
Let G be the matrix with gij as its (i, j)-th entry, and let g ij be the (i, j)-th entry of
G−1 . Similarly, define geαβ to be the inverse of geαβ . Show that:
X X
g ik hkj dui ⊗ duj = geαγ e
hγβ dv α ⊗ dv β .
i,j α,β
84 3. Tensors and Differential Forms
In the previous section we have seen how to express k-linear maps over tangent
vectors using tensor notations. To deal with the above alternating tensors, it is more
elegant and concise to use alternating tensors, or wedge products that we are going to
learn in this section.
3.4.1. Wedge Product on Vector Spaces. Let’s start from the easiest case. Sup-
pose V is a finite dimensional vector space and V ∗ is the dual space of V . Given any two
elements T, S ∈ V ∗ , the tensor product T ⊗ S is a map given by:
(T ⊗ S)(X, Y ) = T (X) S(Y )
for any X, Y ∈ V . The wedge product T ∧ S, where T, S ∈ V ∗ , is a bilinear map defined
by:
T ∧ S := T ⊗ S − S ⊗ T
meaning that for any X, Y ∈ V , we have:
(T ∧ S)(X, Y ) = (T ⊗ S)(X, Y ) − (S ⊗ T )(X, Y )
= T (X) S(Y ) − S(X) T (Y )
It is easy to note that T ∧ S = −S ∧ T .
Take the cross product in R3 as an example. Write the cross product as a bilinear
map ω(a, b) := a × b. It is a (2, 1)-tensor on R3 which can be represented as:
ω = e1 ⊗ e2 ⊗ e3 − e2 ⊗ e1 ⊗ e3
+ e2 ⊗ e3 ⊗ e1 − e3 ⊗ e2 ⊗ e1
+ e3 ⊗ e1 ⊗ e2 − e1 ⊗ e3 ⊗ e2
Now using the wedge product notations, we can express ω as:
ω = (e1 ∧ e2 ) ⊗ e3 + (e2 ∧ e3 ) ⊗ e1 + (e3 ∧ e1 ) ⊗ e2
which is a half shorter than using tensor products alone.
3.4. Wedge Products 85
Now given three elements T1 , T2 , T3 ∈ V ∗ , one can also form a triple wedge product
T1 ∧ T2 ∧ T3 which is a (3, 0)-tensor so that switching any pair of Ti and Tj (with i 6= j)
will give a negative sign. For instance:
T1 ∧ T2 ∧ T3 = −T2 ∧ T1 ∧ T3 and T1 ∧ T2 ∧ T3 = −T3 ∧ T2 ∧ T1 .
It can be defined in a precise way as:
T1 ∧ T2 ∧ T3 := T1 ⊗ T2 ⊗ T3 − T1 ⊗ T3 ⊗ T2
+ T2 ⊗ T3 ⊗ T1 − T2 ⊗ T1 ⊗ T3
+ T3 ⊗ T1 ⊗ T2 − T3 ⊗ T2 ⊗ T1
Exercise 3.29. Verify that the above definition of triple wedge product will result in
T1 ∧ T2 ∧ T3 = −T3 ∧ T2 ∧ T1 .
The above exercise shows that we can decompose (12345)(31) into a product of three
transpositions (32), (15) and (14). In fact, any element in Sn can be decomposed this
way. Here we state a standard theorem in elementary group theory:
X
sgn(σ)Tσ(1) ⊗ Tσ(2) ⊗ Tσ(3)
σ∈S3
= T1 ⊗ T2 ⊗ T3 σ = id
− T2 ⊗ T1 ⊗ T3 σ = (12)
− T3 ⊗ T2 ⊗ T1 σ = (13)
− T1 ⊗ T3 ⊗ T2 σ = (23)
+ T2 ⊗ T3 ⊗ T1 σ = (123) = (13)(12)
+ T3 ⊗ T1 ⊗ T2 σ = (132) = (12)(13)
In general, we define:
Definition 3.30 (Wedge Product). Let V be a finite dimensional vector space, and V ∗
be the dual space of V . Then, given any T1 , . . . , Tk ∈ V ∗ , we define their k-th wedge
product by: X
T1 ∧ · · · ∧ Tk := sgn(σ) Tσ(1) ⊗ . . . ⊗ Tσ(k)
σ∈Sk
where Sk is the permutation group of {1, . . . , k}. The vector space spanned by T1 ∧
· · · ∧ Tk ’s (where T1 , . . . , Tk ∈ V ∗ ) is denoted by ∧k V ∗ .
If we switch any pair of the Ti ’s, then the wedge product differs by a minus sign.
To show this, let τ ∈ Sk be a transposition, then for any σ ∈ Sk , we have sgn(σ ◦ τ ) =
−sgn(σ). Therefore, we get:
X
Tτ (1) ∧ · · · ∧ Tτ (k) = sgn(σ)Tσ(τ (1)) ⊗ . . . ⊗ Tσ(τ (k))
σ∈Sk
X
=− sgn(σ ◦ τ )Tσ◦τ (1) ⊗ . . . ⊗ Tσ◦τ (k)
σ∈Sk
X
=− sgn(σ 0 )Tσ0 (1) ⊗ . . . ⊗ Tσ0 τ (k) (where σ 0 := σ ◦ τ )
σ∈Sk
= −T1 ∧ · · · ∧ Tk .
The last step follows from the fact that σ 7→ σ ◦ τ is a bijection between Sk and itself.
Exercise 3.34. Let {ei }ni=1 be a basis for a vector space V , and {ei }ni=1 be the
corresponding dual basis for V ∗ . Show that:
ei1 ∧ · · · ∧ eik (ej1 , . . . , ejk ) = δi1 j1 · · · δik jk .
Proposition 3.33. Let V be a finite dimensional vector space, and V ∗ be the dual space
of V . Given any σ ∈ ∧k V ∗ and η ∈ ∧r V ∗ , we have:
(3.12) σ ∧ η = (−1)kr η ∧ σ.
Clearly from (3.12), any ω ∈ ∧even V ∗ commutes with any σ ∈ ∧k V ∗ .
where aij are real constants. Then, the wedge product of all ωi ’s are given by:
n
X n
X n
X
ω1 ∧ · · · ∧ ωn = a1j1 ej1 ∧ a2j2 ej2 ∧ · · · ∧ anjn ejn
j1 =1 j2 =1 jn =1
X
j1 jn
= a1j1 a2j2 . . . anjn e ∧ · · · ∧ e
j1 ,...,jn distinct
X
= a1σ(1) a2σ(2) . . . anσ(n) eσ(1) ∧ · · · ∧ eσ(n)
σ∈Sn
Proposition 3.34. Let V ∗ be the dual space of a vector space V of dimension n, and let
{ei }ni=1 be a basis for V , and {ei }ni=1 be the corresponding dual basis for V ∗ . Given any n
Xn
elements ωi = aij ej ∈ V ∗ , we have:
j=1
ω1 ∧ · · · ∧ ωn = (det A) e1 ∧ · · · ∧ en ,
where A is the n × n matrix whose (i, j)-th entry is aij .
where αi are smooth functions locally defined near p. Since the based point p can usually
be understood from the context, we usually denote α by simply:
n
X
α= αi dui .
i=1
Since Tp∗ M is a finite dimensional vector space, we can consider the wedge products
of its elements. A differential k-form ω on a smooth manifold M is a map which assigns
each point p ∈ M to an element in ∧k Tp∗ M . Precisely:
Remark 3.36. It is a convention to denote ∧0 T ∗ M := C ∞ (M, R), the vector space of all
smooth scalar functions defined on M .
We will mostly deal with differential k-forms that are smooth. Therefore, we will
very often call a smooth differential k-form simply by a differential k-form, or even simpler,
a k-form. As we will see in the next section, the language of differential forms will unify
and generalize the curl, grad and div in Multivariable Calculus and Physics courses.
90 3. Tensors and Differential Forms
| ∧ ·{z
Compute ω · · ∧ ω}.
n times
Exercise 3.40. Let (u1 , . . . , un ) and (v1 , . . . , vn ) be two local coordinates of a smooth
manifold M . Show that:
∂(u1 , . . . , un ) 1
du1 ∧ · · · ∧ dun = det dv ∧ · · · ∧ dv n .
∂(v1 , . . . , vn )
Although (3.13) involves local coordinates, it can be easily shown that df is independent
of local coordinates. Suppose (v1 , . . . , vn ) is another local coordinates of M which overlap
with (u1 , . . . , un ). By the chain rule, we have:
n
∂f X ∂f ∂vk
=
∂ui ∂vk ∂ui
k=1
n
X ∂vk i
dv k = du
i=1
∂ui
which combine to give:
n n Xn n
X ∂f X ∂f ∂vk i X ∂f
dui = du = dv k .
i=1
∂ui i=1
∂vk ∂ui ∂vk
k=1 k=1
Therefore, if f is smooth on M then df is a smooth 1-form on M . The components of df
∂f
are ∂u i
’s, and so df is analogous to ∇f in Multivariable Calculus. Note that as long as
f is C ∞ just in an open set U ⊂ M , we can also define df locally on U since (3.13) is a
local expression.
Exterior derivatives can also be defined on differential forms of higher degrees. Let
α ∈ ∧1 T ∗ M , which can be locally written as:
Xn
α= αi dui
i=1
where αi ’s are smooth functions locally defined in a local coordinate chart. Then, we
define:
n n X n
X X ∂αi j
(3.14) dα := dαi ∧ dui = du ∧ dui .
i=1 i=1 j=1
∂u j
Using the fact that duj ∧ dui = −dui ∧ duj and dui ∧ dui = 0, we can also express dα as:
X ∂αi ∂αj
dα = − duj ∧ dui .
∂uj ∂ui
1≤j<i≤n
92 3. Tensors and Differential Forms
Example 3.38. Take M = R3 as an example, and let (x, y, z) be the (usual) coordinates
of R3 , then given any 1-form α = P dx + Q dy + R dz (which is analogous to the vector
field P î + Qĵ + Rk̂), we have:
dα = dP ∧ dx + dQ ∧ dy + dR ∧ dz
∂P ∂P ∂P ∂Q ∂Q ∂Q
= dx + dy + dz ∧ dx + dx + dy + dz ∧ dy
∂x ∂y ∂z ∂x ∂y ∂z
∂R ∂R ∂R
+ dx + dy + dz ∧ dz
∂x ∂y ∂z
∂P ∂P ∂Q ∂Q
= dy ∧ dx + dz ∧ dx + dx ∧ dy + dz ∧ dy
∂y ∂z ∂x ∂z
∂R ∂R
+ dx ∧ dz + dy ∧ dz
∂x ∂y
∂Q ∂P ∂R ∂P ∂R ∂Q
= − dx ∧ dy − − dz ∧ dx + − dy ∧ dz
∂x ∂y ∂x ∂z ∂y ∂z
which is analogous to ∇ × P î + Qĵ + Rk̂ by declaring the correspondence {î, ĵ, k̂}
with {dy ∧ dz, dz ∧ dx, dx ∧ dy}.
One can check that the definition of dα stated in (3.14) is independent of local
coordinates. On general k-forms, the exterior derivatives are defined in a similar way as:
we define:
n
X
(3.15) dω := dωj1 ···jk ∧ duj1 ∧ · · · ∧ dujk
j1 ,··· ,jk =1
n n
X X ∂ωj1 ···jk i
= du ∧ duj1 ∧ · · · ∧ dujk
j1 ,··· ,jk =1 i=1
∂ui
In particular, if ω is an n-form (where n = dim V ), we have dω = 0.
Exercise 3.42. Show that dω defined as in (3.15) does not depend on the choice of
local coordinates.
Example 3.40. Consider R2 equipped with polar coordinates (r, θ). Consider the 1-form:
ω = (r sin θ) dr.
Then, we have
∂(r sin θ) ∂(r sin θ)
dω = dr ∧ dr + dθ ∧ dr
∂r ∂θ
= 0 + (r cos θ) dθ ∧ dr
= −(r cos θ) dr ∧ dθ.
3.5. Exterior Derivatives 93
Proposition 3.41. For any k-forms ω and η, and any smooth scalar function f , we have
the following:
(1) d(ω + η) = dω + dη
(2) d(f ω) = df ∧ ω + f dω
Proof. (1) is easy to prove (left as an exercise for readers). To prove (2), we consider
n
X
local coordinates (u1 , . . . , un ) and let ω = ωj1 ···jk duj1 ∧ · · · ∧ dujk . Then, we
j1 ,...,jk =1
have:
n n
X X ∂
d(f ω) = (f ωj1 ···jk ) dui ∧ duj1 ∧ · · · ∧ dujk
j1 ,...,jk =1 i=1
∂u i
n n
X X ∂f ∂ωj1 ···jk
= ωj1 ···jk + f dui ∧ duj1 ∧ · · · ∧ dujk
j1 ,...,jk =1 i=1
∂u i ∂u i
n
! n
X ∂f X
= dui ∧ ωj1 ···jk duj1 ∧ · · · ∧ dujk
i=1
∂u i j ,...,j =1
1 k
n n
X X ∂ωj1 ···jk i
+f du ∧ duj1 ∧ · · · ∧ dujk
j1 ,··· ,jk =1 i=1
∂ui
as desired.
Identity (2) in Proposition 3.41 can be regarded as a kind of product rule. Given a
k-form α and a r-form β, the general product rule for exterior derivative is stated as:
Exercise 3.45. Prove Proposition 3.42. Based on your proof, explain briefly why
the product rule does not involve any factor of (−1)r .
An crucial property of exterior derivatives is that the composition is zero. For instance,
given a smooth scalar function f (x, y, z) defined on R3 , we have:
∂f ∂f ∂f
df = dx + dy + dz.
∂x ∂y ∂z
Taking exterior derivative one more time, we get:
∂ ∂f ∂ ∂f ∂ ∂f
d(df ) = dx + dy + dz ∧ dx
∂x ∂x ∂y ∂x ∂z ∂x
∂ ∂f ∂ ∂f ∂ ∂f
+ dx + dy + dz ∧ dy
∂x ∂y ∂y ∂y ∂z ∂y
∂ ∂f ∂ ∂f ∂ ∂f
+ dx + dy + dz ∧ dz
∂x ∂z ∂y ∂z ∂z ∂z
∂ ∂f ∂ ∂f ∂ ∂f ∂ ∂f
= − dx ∧ dy + − dz ∧ dx
∂x ∂y ∂y ∂x ∂z ∂x ∂x ∂z
∂ ∂f ∂ ∂f
+ − dy ∧ dz
∂y ∂z ∂z ∂y
Since partial derivatives commute, we get d(df ) = 0, or in short d2 f = 0, for any scalar
function f . The fact that d2 = 0 is generally true on smooth differential forms, not only
for scalar functions. Precisely, we have:
n
X
Proof. Let ω = ωj1 ···jk duj1 ∧ · · · ∧ dujk , then:
j1 ,...,jk =1
n n
X X ∂ωj 1 ···jk
dω = dui ∧ duj1 ∧ · · · ∧ dujk .
j1 ,··· ,jk =1 i=1
∂ui
n n
X X ∂ωj 1 ···jk
d2 ω = d dui ∧ duj1 ∧ · · · ∧ dujk
j1 ,··· ,jk =1 i=1
∂ui
n n X
n
X X ∂ 2 ωj 1 ...jk
= dul ∧ dui ∧ duj1 ∧ · · · ∧ dujk
j1 ,··· ,jk =1 i=1 l=1
∂ul ∂ui
n
X ∂ 2 ωj1 ...jk l
For each fixed k-tuple (j1 , . . . , jk ), the term du ∧ dui can be rewritten as:
∂ul ∂ui
i,l=1
Definition 3.44 (Exact and Closed Forms). Let ω be a smooth k-form defined on a
smooth manifold M , then we say:
• ω is exact if there exists a (k − 1)-form η defined on M such that ω = dη;
• ω is closed if dω = 0.
Remark 3.45. By the fact that d2 = 0 (Proposition 3.43), it is clear that every exact form
is a closed form (but not vice versa).
The list below showcases the corresponding concepts of exact/closed forms in Multi-
variable Calculus.
Differential Form on R3 Multivariable Calculus
exact 1-form conservative vector field
closed 1-form curl-less vector field
exact 2-form solenoidal vector field
closed 2-form divergence-less vector field
Example 3.46. On R3 , the 1-form:
α = yz dx + zx dy + xy dz
is exact since α = df where f (x, y, z) = xyz. By Proposition 3.43, we immediately get
dα = d(df ) = 0, so α is a closed form. One can also verify this directly:
dα = (z dy + y dz) ∧ dx + (z dx + x dz) ∧ dy + (y dx + x dy) ∧ dz
= (z − z) dx ∧ dy + (y − y) dz ∧ dx + (x − x) dy ∧ dz = 0.
Example 3.47. The 1-form:
y x
α := − dx + 2 dy
x2 + y 2 x + y2
defined on R2 \{(0, 0)} is closed:
∂ y ∂ x
dα = − 2 dy ∧ dx + dx ∧ dy
∂y x + y2 ∂x x2 + y 2
y 2 − x2 y 2 − x2
= 2 2 2
dy ∧ dx + 2 dx ∧ dy
(x + y ) (x + y 2 )2
=0
as dx ∧ dy = −dy ∧ dx. However, we will later see that α is not exact.
y
Note that even though we have α = df where f (x, y) = tan−1 , such an f is NOT
x
smooth on R2 \{(0, 0)}. In order to claim α is exact, we require such an f to be smooth
on the domain of α.
3.5. Exterior Derivatives 97
Exercise 3.48. The purpose of this exercise is to show that any closed 1-form ω on
R3 must be exact. Let
ω = P (x, y, z) dx + Q(x, y, z) dy + R(x, y, z) dz
be a closed 1-form on R3 . Define f : R3 → R by:
Z t=1
f (x, y, z) = (xP (tx, ty, tz) + yQ(tx, ty, tz) + zR(tx, ty, tz)) dt
t=0
Show that ω = df . Point out exactly where you have used the fact that dω = 0.
3.5.4. Pull-Back of Tensors. Let’s first begin by reviewing the push-forward and
pull-back of tangent and cotangent vectors. Given a smooth map Φ : M → N between
two smooth manifolds M m and N n , its tangent map Φ∗ takes a tangent vector in
Tp M to a tangent vector in TΦ(p) N . If we let F (u1 , . . . , um ) be local coordinates of M ,
G(v1 , . . . , vn ) be local coordinates of N and express the map Φ locally as:
(v1 , . . . , vn ) = G−1 ◦ Φ ◦ F (u1 , . . . , um ),
∂
then Φ∗ acts on the basis vectors by:
∂ui
∂ ∂Φ X ∂vj ∂
Φ∗ = = .
∂ui ∂ui j
∂ui ∂vj
The tangent map Φ∗ is also commonly called the push-forward map. It is important to
∂v
note that the vj ’s in the partial derivatves ∂uji can sometimes cause confusion if we talk
about the push-forwards of two different smooth maps Φ : M → N and Ψ : M → N .
Even with the same input (u1 , . . . , um ), the output Φ(u1 , . . . , um ) and Ψ(u1 , . . . , um ) are
generally different and have different vj -coordinates. To avoid this confusion, it is best
to write:
X
∂ ∂(vj ◦ Φ) ∂
Φ∗ =
∂ui j
∂ui ∂vj
X
∂ ∂(vj ◦ Ψ) ∂
Ψ∗ =
∂ui j
∂ui ∂vj
∂v
Here each vj in the partial derivatives ∂uji are considered to be a locally defined function
taking a point p ∈ N to its vj -coordinate.
For cotangent vectors (i.e. 1-forms), we talk about pull-back instead. According to
Definition 3.14, Φ∗ takes a cotangent vector in TΦ(p)
∗
N to a cotangent vector in Tp∗ M ,
defined as follows:
Φ∗ (dv i )(X) = dv i (Φ∗ X) for any X ∈ Tp M .
In terms of local coordinates, it is given by:
X ∂(vi ◦ Φ)
Φ∗ (dv i ) = duj .
j
∂u j
To avoid confusion, we use (x1 , x2 ) to label the coordinates of the domain R2 , and use
(y1 , y2 , y3 ) to denote the coordinates of the codomain R3 . Then, we have:
∗ 1 ∂ 1 ∂ 1 ∂Φ
Φ (dy ) = dy Φ∗ = dy
∂x1 ∂x1 ∂x1
1 ∂(y1 ◦ Φ) ∂ ∂(y2 ◦ Φ) ∂ ∂(y3 ◦ Φ) ∂
= dy + +
∂x1 ∂y1 ∂x1 ∂y2 ∂x1 ∂y3
∂(y1 ◦ Φ) ∂ x1 +x2
= = e = ex1 +x2 .
∂x1 ∂x1
Similarly, we have:
∂ ∂(y1 ◦ Φ) ∂ x1 +x2
Φ∗ (dy 1 ) = = e = ex1 +x2 .
∂x2 ∂x2 ∂x2
Therefore, Φ∗ (dy 1 ) = ex1 +x2 dx1 + ex1 +x2 dx2 = ex1 +x2 (dx1 + dx2 ). We leave it as an
exercise for readers to verify that:
Φ∗ (dy 2 ) = 2x1 x2 cos(x21 x2 ) dx1 + x21 cos(x21 x2 ) dx2
Φ∗ (dy 3 ) = dx1
a local parametrization of Σ, then ι∗ (dx) should be in terms of du and dv, but not dx, dy
and dz. Precisely, we have:
∂F ∂ι ∂(ι ◦ F ) ∂F
ι∗ = := =
∂u ∂u ∂u ∂u
∂F ∂F ∂F
ι∗ (dx) = dx ι∗ = dx
∂u ∂u ∂u
∂x ∂ ∂y ∂ ∂z ∂
= dx + +
∂u ∂x ∂u ∂y ∂u ∂z
∂x
= .
∂u
∂F ∂x
Similarly, we also have ι∗ (dx) = , and hence:
∂v ∂v
∂x ∂x
ι∗ (dx) = du + dv.
∂u ∂v
As a result, we have:
∂F ∂F ∂F ∂F ∂F ∂F ∂F ∂F
ι∗ ω = · du ⊗ du + · du ⊗ dv + · dv ⊗ du + · dv ⊗ dv,
∂u ∂u ∂u ∂v ∂v ∂u ∂v ∂v
Exercise 3.49. Let the unit sphere S2 be locally parametrized by spherical coordi-
nates (θ, ϕ). Consider the (2, 0)-tensor on R3 :
ω = x dy ⊗ dz
∗
Express the pull-back ι ω in terms of (θ, ϕ).
One can derive a general formula (which you do not need to remember in practice)
for the local expression of pull-backs. Consider local coordinates {ui } for M and {vi } for
N , and write (v1 , . . . , vn ) = Φ(u1 , . . . , um ) and
n
X
T = Ti1 ···ik (v1 , . . . , vn ) dv i1 ⊗ · · · ⊗ dv ik .
i1 ,...,ik =1
100 3. Tensors and Differential Forms
n m
X X ∂vi1 ∂vik
= Ti1 ···ik (Φ(u1 , . . . , um )) ··· duj1 ⊗ · · · ⊗ dujk .
i1 ,...,ik =1 j1 ,...,jk =1
∂uj1 ∂ujk
In view of Ti1 ···ik (v1 , . . . , vn ) = Ti1 ···ik (Φ(u1 , . . . , um )) and the above local expression,
we define
Φ∗ f := f ◦ Φ
for any scalar function of f . Using this notation, we then have Φ∗ (f T ) = (Φ∗ f ) Φ∗ T for
any scalar function f and (k, 0)-tensor T .
Exercise 3.50. Let Φ : M → N be a smooth map between smooth manifolds M and
N , f be a smooth scalar function defined on N . Show that
Φ∗ (df ) = d(Φ∗ f ).
In particular, if (v1 , . . . , vn ) are local coordinates of N , we have Φ∗ (dv j ) = d(Φ∗ v j ).
Example 3.53. Using the result from Exercise 3.50, one can compute the pull-back
by inclusion map ι : Σ → R3 for regular surfaces Σ in R3 . Suppose F (u, v) is a local
parametrization of Σ, then:
ι∗ (dx) = d(ι∗ x) = d(x ◦ ι).
Although x ◦ ι and x (as a coordinate function) have the same output, their domains are
different! Namely, x ◦ ι : Σ → R while x : R3 → R. Therefore, when computing d(x ◦ ι),
one should express it in terms of local coordinates (u, v) of Σ:
∂(x ◦ ι) ∂(x ◦ ι) ∂x ∂x
d(x ◦ ι) = du + dv = du + dv.
∂u ∂v ∂u ∂v
Recall that the tangent maps (i.e. push-forwards) acting on tangent vectors satisfy
the chain rule: if Φ : M → N and Ψ : N → P are smooth maps between smooth
manifolds, then we have (Ψ ◦ Φ)∗ = Ψ∗ ◦ Φ∗ . It is easy to extend the chain rule to
(k, 0)-tensors:
Exercise 3.52. Denote idM and idT M to be the identity maps of a smooth manifold
∗
M and its tangent bundle respectively. Show that (idM ) = idT M . Hence, show that
if M and N are diffeomorphic, then for k ≥ 1 the vector spaces of (k, 0)-tensors
⊗k T ∗ M and ⊗k T ∗ N are isomorphic.
3.5.5. Pull-Back of Differential Forms. By linearity of the pull-back map, and the
fact that differential forms are linear combinations of tensors, the pull-back map acts on
differential forms by the following way:
Φ∗ (T1 ∧ · · · ∧ Tk ) = Φ∗ T1 ∧ · · · ∧ Φ∗ Tk
for any 1-forms T1 , . . . , Tk .
Example 3.55. Consider the map Φ : R2 → R2 given by:
Φ(x1 , x2 ) = (x21 − x2 , x32 ).
| {z }
(y1 ,y2 )
Proof. Proceed as in the derivation of (3.16) by simply replacing all tensor products by
wedge products, we get:
m
∗ i1 ik
X ∂vi1 ∂vik j1 jk
Φ (dv ∧ · · · ∧ dv ) = ··· du ∧ · · · ∧ du
j1 ,...,jk =1
∂uj1 ∂ujk
m
X ∂vi1 ∂vik j1 jk
= ··· du ∧ · · · ∧ du
j ,...,j =1
∂uj1 ∂ujk
1 k
j1 ,...,jk distinct
102 3. Tensors and Differential Forms
The second equality follows from the fact that duj1 ∧ · · · ∧ dujk = 0 if {j1 , . . . , jk } are not
all distinct. Each k-tuples (j1 , . . . , jk ) with distinct ji ’s can be obtained by permuting a
strictly increasing sequence of j’s. Precisely, we have:
{(j1 , . . . , jk ) : 1 ≤ j1 , . . . , jk ≤ n and j1 , . . . , jk are all distinct}
[
= {(jσ(1) , . . . , jσ(k) ) : 1 ≤ j1 < j2 < . . . < jk ≤ n}
σ∈Sk
Therefore, we get:
Φ∗ (dv i1 ∧ · · · ∧ dv ik )
!
X X ∂vi1 ∂vik
= ··· dujσ(1) ∧ · · · ∧ dujσ(k)
∂ujσ(1) ∂ujσ(k)
1≤j1 <...<jk ≤m σ∈Sk
X X ∂vi1 ∂vik
= sgn(σ) ··· duj1 ∧ · · · ∧ dujk
∂ujσ(1) ∂ujσ(k)
1≤j1 <...<jk ≤m σ∈Sk
X ∂vi1 ∂vik ∂vip
By observing that sgn(σ) ··· is the determinant of ,
∂ujσ(1) ∂ujσ(k) ∂ujq 1≤p,q≤k
σ∈Sk
the desired result (3.18) follows easily.
The second result (3.19) follows directly from (3.18). In case of dim M = dim N = n
and k = n, the only possible strictly increasing sequence 1 ≤ j1 < . . . < jn ≤ n is
(j1 , . . . , jn ) = (1, 2, . . . , n).
Proposition 3.57. Let Φ : M → N be a smooth map between two smooth manifolds. For
any ω ∈ ∧k T ∗ N , we have:
(3.20) Φ∗ (dω) = d(Φ∗ ω).
To be precise, we say Φ∗ (dN ω) = dM (Φ∗ ω), where dN : ∧k T ∗ N → ∧k+1 T ∗ N and
dM : ∧k T ∗ M → ∧k+1 T ∗ M are the exterior derivatives on N and M respectively.
Proof. Let {uj } and {vi } be local coordinates of M and N respectively. By linearity, it
suffices to prove (3.20) for the case ω = f dv i1 ∧ · · · ∧ dv ik where f is a locally defined
scalar function. The proof follows from computing both LHS and RHS of (3.20):
dω = df ∧ dv i1 ∧ · · · ∧ dv ik
Φ∗ (dω) = Φ∗ (df ) ∧ Φ∗ (dv i1 ) ∧ · · · ∧ Φ∗ (dv ik )
= d(Φ∗ f ) ∧ d(Φ∗ v j1 ) ∧ · · · ∧ d(Φ∗ v jk ).
Here we have used Exercise 3.50. On the other hand, we have:
Φ∗ ω = (Φ∗ f ) Φ∗ (dv j1 ) ∧ · · · ∧ Φ∗ (dv jk )
= (Φ∗ f ) d(Φ∗ v i1 ) ∧ · · · ∧ d(Φ∗ v ik )
d(Φ∗ ω) = d(Φ∗ f ) ∧ d(Φ∗ v i1 ) ∧ · · · ∧ d(Φ∗ v ik )
+ Φ∗ f d d(Φ∗ v i1 ) ∧ · · · ∧ d(Φ∗ v ik )
Since d2 = 0, each of d(Φ∗ v iq ) is a closed 1-form. By Proposition 3.42 (product rule) and
induction, we can conclude that:
d d(Φ∗ v i1 ) ∧ · · · ∧ d(Φ∗ v ik ) = 0
Exercise 3.53. Show that the pull-back of any closed form is closed, and the pull-
back of any exact form is exact.
ω = z dx ∧ dy. Compute dω, Φ∗ (dω), Φ∗ ω and d(Φ∗ ω), and verify they satisfy
Proposition 3.57.
Now consider a regular surface M ⊂ R3 . Suppose F (u, v) = (x(u, v), y(u, v), z(u, v))
is a smooth local parametrization of M . Consider a vector G = βx î + βy ĵ + βz k̂ on R3
and its corresponding 2-form on R3 :
β = βx dy ∧ dz + βy dz ∧ dx + βz dx ∧ dy.
Denote ι : M → R the inclusion map. The induced 2-form ι∗ β on M is in fact related to
3
Recall that the classical Stokes’ Theorem is related to line integrals of a curve
and surface flux of a vector field. Based on the above discussion, we see that Stokes’
Theorem can be restated in terms of differential forms. Consider the 1-form α =
αx dx + αy dy + αz dz and its corresponding vector field F = αx î + αy ĵ + αz k̂. We have
already discussed that the 2-form dα corresponds to the vector field ∇×F . Therefore, the
surface flux of the vector field ∇ × F through M can be expressed in terms of differential
forms as: ZZ ZZ ZZ
∗
(∇ × F ) · ν dS = ι (dα) = d(ι∗ α).
M M M
3.5. Exterior Derivatives 105
If C is the boundary curve of M , then from our previous discussion we can write:
Z Z
F · dl = ι∗ α.
C C
Due to this elegant way (although not very practical for physicists and engineers) of
expressing Stokes’ Theorem, we often denote the boundary of a surface M as ∂M , then
the classical Stokes’ Theorem can be expressed as:
Z ZZ
α= dα.
∂M M
Using differential forms, one can also express Divergence Theorem in Multivariable
Calculus in a similar way as above. Let D be a solid region in R3 and ∂D be the boundary
surface of D. Divergence Theorem in MATH 2023 asserts that:
ZZ ZZZ
G · ν dS = ∇ · G dV,
∂D D
ZZ
where G = βx î + βy ĵ + βz k̂. As discussed before, the LHS is β where β =
∂D
βx dy ∧ dz + βy dz ∧ dx + βz dx ∧ dy. We have seen that:
∂βx ∂βy ∂βz
dβ = + + dx ∧ dy ∧ dz,
∂x ∂y ∂z
which is (almost) the same as:
∂βx ∂βy ∂βz
∇ · G dV = + + dxdydz.
∂x ∂y ∂z
ZZZ
Hence, the RHS of Divergence Theorem can be expressed as dβ; and therefore we
D
can rewrite Divergence Theorem as:
ZZ ZZZ
β= dβ.
∂D D
Again, the same expression! Stokes’ and Divergence Theorems can therefore be unified.
Green’s Theorem can also be unified with Stokes’ and Divergence Theorems as well. Try
the exercise below:
Exercise 3.55. Let C be a simple closed smooth curve in R2 and R be the region
enclosed by C in R2 . Given a smooth vector field F = P î + Qĵ on R2 , Green’s
Theorem asserts that:
Z ZZ
∂Q ∂P
F · dl = − dxdy.
C R ∂x ∂y
Express Green’s Theorem using the language of differential forms.
106 3. Tensors and Differential Forms
3.5.7. Differential Forms and Maxwell’s Equations. The four Maxwell’s equa-
tions are a set of partial differential equations that form the foundation of electromag-
netism. Denote the components of the electric field e, magnetic field B, and current
density ĵ by
E = Ex î + Ey ĵ + Ez k̂
B = Bx î + By ĵ + Bz k̂
J = Jx î + Jy ĵ + Jz k̂
∇·E=ρ ∇·B=0
∂B ∂E
∇×E=− ∇×B=J+
∂t ∂t
These four equations can be rewritten using differential forms in a very elegant way.
Consider R4 with coordinates (t, x, y, z), which is also denoted as (x0 , x1 , x2 , x3 ) in this
problem. First we introduce the Minkowski Hodge-star operator ∗ on R4 , which is a linear
map taking p-forms on R4 to (4 − p)-forms on R4 . In particular, for 2-forms ω = dxi ∧ dxj
(where i, j = 0, 1, 2, 3 and i 6= j), we define ∗ω to be the unique 2-form on R4 such that:
(
dt ∧ dx ∧ dy ∧ dz if i, j 6= 0
ω ∧ ∗ω =
−dt ∧ dx ∧ dy ∧ dz otherwise.
E = Ex dx + Ey dy + Ez dz
B = Bx dy ∧ dz + By dz ∧ dx + Bz dx ∧ dy
J = −(Jx dy ∧ dz + Jy dz ∧ dx + Jz dx ∧ dy) ∧ dt + ρ dx ∧ dy ∧ dz
Note that Ei ’s and Bj ’s may depend on t although there is no dt above. Define the
2-form:
F := B + E ∧ dt.
Exercise 3.57. Show that the four Maxwell’s equations can be rewritten in an
elegant way as:
dF = 0
d (∗F ) = J
where d is the exterior derivative on R4 .
3.5. Exterior Derivatives 107
Exercise 3.58. Guess the definition of Lie derivatives of a general (p, q)-tensor along
a vector field X. Check any standard textbook to see if your guess is right.
Remark 3.58. On a regular surface M in R3 with the first fundamental form denoted by
g, if X is a vector field on M such that LX g = 0, then we call X to be a Killing vector
field. The geometric meaning of such an X is that g is invariant when M moves along
the vector field X, or equivalently, g is symmetric in the direction of X. This concept
of Killing vector fields can be generalize to Riemannian manifolds and is important in
Differential Geometry and General Relativity, whenever symmetry plays an important
role.
Pn ∂
Exercise 3.60. Let X = i=1 X i ∂u i
be a vector field on a manifold M with local
coordinates (u1 , . . . , un ). Derive the local expression of:
iX (duj1 ∧ duj2 ∧ · · · ∧ dujk )
where 1 ≤ j1 < j2 < · · · < jk ≤ n.
Now we are ready to present a beautiful and elegant formula due to Elie Cartan:
Proposition 3.61 (Cartan’s Magic Formula). Let X be a smooth vector field on a manifold
M , then for any differential k-form ω, we have:
(3.22) LX ω = iX (dω) + d(iX ω)
Proof. The proof is by induction on k, the degree of ω. We first show that (3.22) holds
for 1-forms.
Pn
Consider ω = j=1 ωj duj , we have already computed in (3.11) that:
n
∂X i
X ∂ωj
LX ω = Xi + ωi duj .
i,j=1
∂ui ∂uj
Moreover, we have
n
X
iX ω = ω(X) = X j ωj
j=1
n
∂X j
X ∂ωj
d(iX ω) = ωj + X j dui
i,j=1
∂ui ∂ui
Exercise 3.61. Complete the above inductive proof. [Note: the proof is somewhat
algebraic.]
Exercise 3.62. Show that if ω is closed, then LX ω is exact for any vector field X.
3.5. Exterior Derivatives 109
Readers interested in the proof may consult [Lee13, P.370, Proposition 14.32].
Chapter 4
Generalized Stokes’
Theorem
4.1.1. Smooth Functions on Upper-Half Spaces. From now on, we denote Rn+ :=
{(u1 , . . . , un ) ∈ Rn : un ≥ 0} which is the upper-half space of Rn . Under the subspace
topology, we say a subset V ⊂ Rn+ is open in Rn+ if there exists a set Ve ⊂ Rn open in Rn
such that V = Ve ∩ Rn+ . It is intuitively clear that if V ⊂ Rn+ is disjoint from the subspace
{un = 0} of Rn , then V is open in Rn+ if and only if V is open in Rn .
Now consider a set V ⊂ Rn+ which is open in Rn+ and that V ∩ {un = 0} 6= ∅. We
need to first develop a notion of differentiability for functions such an V as their domain.
Given a vector-valued function G : V → Rm , then near a point u ∈ V ∩ {un = 0}, we can
only approach u from one side only, namely from directions with positive un -coordinates.
The usual definition of differentiability does not apply at such a point, so we define:
111
112 4. Generalized Stokes’ Theorem
Definition 4.1 (Functions of Class C k on Rn+ ). Let V ⊂ Rn+ be open in Rn+ and that
V ∩ {un = 0} 6= ∅. Consider a vector-valued function G : V → Rm . We say G is C k
(resp. smooth) at u ∈ V ∩ {un = 0} if there exists a C k (resp. smooth) local extension
Ge : Bε (u) → Rm such that G(y)
e = G(y) for any y ∈ Bε (u) ∩ V . Here Bε (u) ⊂ Rn refers
n
to an open ball in R .
If G is C k (resp. smooth) at every u ∈ V (including those points with un > 0), then
we say G is C k (resp. smooth) on V .
Example 4.2. Let V = {(x, y) : y ≥ 0 and x2 + y 2 < 1}, which is an open set in R2+
since V = {(x, y) : x2 + y 2 < 1} ∩ R2+ . Then f (x, y) : V → R defined by f (x, y) =
| {z }
open in R2
p p
x2 y2
1 − − is a smooth function on V since 1 − x2 − y 2 is smoothly on the whole
ball x2 + y 2 < 1.
√
However, the function g : V → R defined by g(x, y) = y is not smooth at every
∂g
point on the y-axis because ∂y → ∞ as y → 0+ . Any extension ge of g will agree with g
∂e
g
on the upper-half plane, and hence will also be true that ∂y → ∞ as y → 0+ , which is
sufficient to argue that such ge is not smooth.
Remark 4.4. In this course, we will call these Fα ’s to be local parametrizations of interior
type, and these Gβ ’s to be local parametrizations of boundary type.
Example 4.5. Consider the solid ball B2 := {x ∈ R2 : |x| ≤ 1}. It can be locally
parametrized using polar coordinates by:
G : (0, 2π) × [0, 1) → B2
G(θ, r) := (1 − r)(cos θ, sin θ)
Note that the domain of G can be regarded as a subset
V := {(θ, r) : θ ∈ (0, 2π) and 0 ≤ r < 1} ⊂ R2+ .
Here we used 1 − r instead of r so that the boundary of B2 has zero r-coordinate, and
the interior of B2 has positive r-coordinate.
Note that the image of G does not cover the whole solid ball B2 . Precisely, the image
of G is B2 \{non-negative x-axis}. In order to complete the proof that B2 is a manifold
with boundary, we cover B2 by two more local parametrizations:
e : (−π, π) × [0, 1) → B2
G
e r) := (1 − r)(cos θ, sin θ)
G(θ,
and also the inclusion map ι : {u ∈ R2 : |u| < 1} → B2 . We need to show that the
transition maps are smooth. There are six possible transition maps:
e −1 ◦ G,
G G−1 ◦ G,
e ι−1 ◦ G, ι−1 ◦ G,
e G−1 ◦ ι, and e −1 ◦ ι.
G
114 4. Generalized Stokes’ Theorem
The first one is given by (we leave it as an exercise for computing these transition maps):
Ge −1 ◦ G : ((0, π) ∪ (π, 2π)) × [0, 1) → ((−π, 0) ∪ (0, π)) × [0, 1)
(
−1 (θ, r) if θ ∈ (0, π)
G ◦ G(θ, r) =
e
(θ − 2π, r) if θ ∈ (π, 2π)
which can be smoothly extended to the domain ((0, π) ∪ (π, 2π)) × (−1, 1). Therefore,
e −1 ◦ G is smooth. The second transition map G−1 ◦ G
G e can be computed and verified to
be smooth in a similar way.
For ι−1 ◦ G, by examining the overlap part of ι and G on B2 , we see that the domain
of the transition map is an open set (0, 2π) × (0, 1) in R2 . On this domain, ι−1 ◦ G is
essentially G, which is clearly smooth. Similar for ι−1 ◦ G.
e
To show G−1 ◦ ι is smooth, we use the Inverse Function Theorem. The domain of
ι ◦ G is (0, 2π) × (0, 1). By writing (x, y) = ι−1 ◦ G(θ, r) = (1 − r)(cos θ, sin θ), we check
−1
From the above example and exercise, we see that verifying a set is a manifold
with boundary may be cumbersome. The following proposition provides us with a very
efficient way to do so.
Proof. We need to construct local parametrizations for the set Σ. Given any point p ∈ Σ,
then by the definition of Σ, we have f (p) > c or f (p) = c.
For the former case f (p) > c, we are going to show that near p there is a local
parametrization of Σ of interior type. Regarding p as a point in the manifold M , there
4.1. Manifolds with Boundary 115
With the help of Proposition 4.6, one can show many sets are manifolds with
boundary by picking a suitable submersion f .
Example 4.8. The n-dimensional ball Bn = {x ∈ Rn : |x| ≤ 1} is an n-manifold with
boundary. To argue this, let f : Rn → R be the function:
2
f (x) = 1 − |x| .
Then Bn = f −1 ([0, ∞)).
The tangent map f∗ is represented by the matrix:
∂f ∂f
[f∗ ] = , ··· , = −2 [x1 , · · · , xn ]
∂x1 ∂xn
which is surjective if and only if (x1 , . . . , xn ) 6= (0, . . . , 0). For any x ∈ f −1 (0), we have
2
|x| = 1 and so in particular x 6= 0. Therefore, f is a submersion at every x ∈ f −1 (0).
By Proposition 4.6, we proved Bn = f −1 ([0, ∞)) is an n-dimensional manifold with
boundary, and the boundary is f −1 (0) = {x ∈ Rn : |x| = 1}, i.e. the unit circle.
116 4. Generalized Stokes’ Theorem
4.2. Orientability
In Multivariable Calculus, we learned (or was told) that Stokes’ Theorem requires the
surface to be orientable, meaning that the unit normal vector ν varies continuously on
the surface. The Möbius strip is an example of non-orientable surface.
Now we are talking about abstract manifolds which may not sit inside any Euclidean
space, and so it does not make sense to define normal vectors to the manifold. Even when
the manifold M is a subset of Rn , if the dimension of the manifold is dim M ≤ n − 2, the
manifold does not have a unique normal vector direction. As such, in order to generalize
the notion of orientability of abstract manifolds, we need to seek a reasonable definition
without using normal vectors.
In this section, we first show that for hypersurfaces M n in Rn+1 , the notion of
orientability using normal vectors is equivalent to another notion using transition maps.
Then, we extend the notion of orientability to abstract manifolds using transition maps.
Let’s explore the above definition a bit in the easy case n = 2. Given a regular surface
M 2 in R3 with a local parametrization (x, y, z) = F (u1 , u2 ) : U → M , one can find a
normal vector to the surface by taking cross product:
∂F ∂F ∂(y, z) ∂(z, x) ∂(x, y)
× = det î + det ĵ + det k̂
∂u1 ∂u2 ∂(u1 , u2 ) ∂(u1 , u2 ) ∂(u1 , u2 )
and hence the unit normal along this direction is given by:
∂(y,z) ∂(z,x) ∂(x,y)
det ∂(u1 ,u2 )
î + det ∂(u 1 ,u2 )
ĵ + det ∂(u1 ,u2 )
k̂
νF = on F (U).
∂(y,z) ∂(z,x) ∂(x,y)
det ∂(u1 ,u2 )
î + det ∂(u 1 ,u2 )
ĵ + det ∂(u1 ,u2 )
k̂
Using the chain rule, we have the following relation between the Jacobian determinants:
∂(∗, ∗∗) ∂(u1 , u2 ) ∂(∗, ∗∗)
det = det det
∂(v1 , v2 ) ∂(v1 , v2 ) ∂(u1 , u2 )
(here ∗ and ∗∗ mean any of the x, y and z) and therefore νF and νG are related by:
det ∂(u 1 ,u2 )
∂(v1 ,v2 )
νG = νF .
det ∂(u 1 ,u2 )
∂(v1 ,v2 )
Therefore, if there is an overlap between local coordinates (u1 , u2 ) and (v1 , v2 ), the unit
normal vectors νF and νG agree with each other on the overlap F (U) ∩ G(V) if and only
∂(u1 , u2 )
if det > 0 (equivalently, det D(F −1 ◦ G) > 0).
∂(v1 , v2 )
118 4. Generalized Stokes’ Theorem
From above, we see that consistency of unit normal vector on different local coor-
dinate charts is closely related to the positivity of the determinants of transition maps.
A consistence choice of unit normal vector ν exists if and only if it is possible to pick
a family of local parametrizations Fα : Uα → M 2 covering the whole M such that
det D(Fβ−1 ◦ Fα ) > 0 on Fα−1 (Fα (Uα ) ∩ Fβ (Uβ )) for any α and β in the family. The
notion of normal vectors makes sense only for hypersurfaces in Rn , while the notion of
transition maps can extend to any abstract manifold.
Note that given two local parametrizations F (u1 , u2 ) and G(v1 , v2 ), it is not always
∂(u1 , u2 )
possible to make sure det > 0 on the overlap even by switching v1 and v2 . It
∂(v1 , v2 )
is because it sometimes happens that the overlap F (U) ∩ G(V) is a disjoint union of
two open sets. If on one open set the determinant is positive, and on another one the
determinant is negative, then switching v1 and v2 cannot make the determinant positive
on both open sets. Let’s illustrate this issue through two contrasting examples: the
cylinder and the Möbius strip:
Example 4.10. The unit cylinder Σ2 in R3 can be covered by two local parametrizations:
Then, the transition map Fe−1 ◦ F is defined on a disconnected domain θ ∈ (0, π) ∪ (π, 2π)
and z ∈ R, and it is given by:
(
−1 (θ, z) if θ ∈ (0, π)
F ◦ F (θ, z) =
e
(θ − 2π, z) if θ ∈ (π, 2π)
D(Fe−1 ◦ F )(θ, z) = I
in either case θ ∈ (0, π) or θ ∈ (π, 2π). Therefore, det D(Fe−1 ◦ F ) > 0 on the overlap.
The unit normal vectors defined using these F and Fe:
∂F ∂F
∂r × ∂θ
νF = ∂F ∂F
on F ((0, 2π) × R)
∂r × ∂θ
∂F ∂F
×
e e
∂e
r ∂ θe
νFe = on Fe((−π, π) × R)
∂F ∂F
×
e e
∂e
r ∂ θe
will agree with each other on the overlap. Therefore, it defines a global continuous unit
normal vector across the whole cylinder.
Example 4.11. The Möbius strip Σ2 in R3 can be covered by two local parametrizations:
In order to compute the transition map Fe−1 ◦ F (u, θ), we need to solve the system of
equations, i.e. find (e e in terms of (u, θ):
u, θ)
!
θ θe
(4.1) 3 + u cos cos θ = 3 + u
e cos cos θe
2 2
!
θ θe
(4.2) 3 + u cos sin θ = 3 + u
e cos sin θe
2 2
θ θe
(4.3) u sin =u
e sin
2 2
By considering (4.1)2 + (4.2)2 , we get:
θ θe
(4.4) u cos =u
e cos
2 2
We leave it as an exercise for readers to check that θ 6= π in order for the system to
be solvable. Therefore, θ ∈ (0, π) ∪ (π, 2π) and so the domain of overlap is a disjoint
union of two open sets.
When θ ∈ (0, π), from (4.3) and (4.4) we can conclude that u e = u and θe = θ.
When θ ∈ (π, 2π), we cannot have θe = θ since θe ∈ (−π, π). However, one can have
e = −u so that (4.3) and (4.4) become:
u
θ θe θ θe
sin = − sin and cos = − cos
2 2 2 2
which implies θe = θ − 2π.
To conclude, we have:
(
e−1 (u, θ) if θ ∈ (0, π)
F ◦ F (u, θ) =
(−u, θ − 2π) if θ ∈ (π, 2π)
Next, we are back to hypersurfaces M n in Rn+1 and prove the equivalence between
consistency of unit normal and positivity of transition maps. To begin, we need the
following result about normal vectors (which is left as an exercise for readers):
Proposition 4.12. Given a smooth hypersurface M n in Rn+1 , the following are equivalent:
(i) M n is orientable;
(ii) There exists a family of local parametrizations Fα : Uα → M covering M such that
for any Fα , Fβ in the family with Fβ (Uβ ) ∩ Fα (Uα ) 6= ∅, we have:
det D(Fα−1 ◦ Fβ ) > 0 on Fβ−1 (Fβ (Uβ ) ∩ Fα (Uα )).
Then on the overlap Fβ−1 (Fα (Uα ) ∩ Fβ (Uβ )), the chain rule asserts that:
By the condition that det D(Fα−1 ◦ Fβ ) > 0, we have νβ = να on the overlap. Define
ν := να on every Fα (Uα ), it is then a continuous unit normal vector globally defined on
M . This proves (i).
Now we show (i) =⇒ (ii). Suppose ν is a continuous unit normal vector defined on
the whole M . Suppose Fα (uα α
1 , . . . , un ) : Uα → M is any family of local parametrizations
that cover the whole M . On every Fα (Uα ), we consider the locally defined unit normal
vector:
Pn+1 ∂(xi+1 ,...,xn+1 ,x1 ,...,xi−1 )
i=1 det ∂(uα α
1 ,...,un )
êi
να = P .
n+1 ∂(xi+1 ,...,xn+1 ,x1 ,...,xi−1 )
i=1 det α
∂(u ,...,u ) α ê i
1 n
n n+1
As a hypersurface M in R , there is only one direction of normal vectors, and so
we have either να = ν or να = −ν on Fα (Uα ). For the latter case, one can modify the
parametrization Fα by switching any pair of uα
i ’s such that να = ν.
After making suitable modification on every Fα , we can assume without loss of
generality that Fα ’s are local parametrizations such that να = ν on every Fα (Uα ). In
particular, on the overlap Fβ−1 (Fα (Uα ) ∩ Fβ (Uβ )), we have να = νβ .
4.2. Orientability 121
det D(Fα−1 ◦ Fβ )
By νβ = να , we conclude that det D(Fα−1 ◦ Fβ ) > 0, proving (ii).
det D(Fα−1 ◦ Fβ )
Remark 4.13. According to Proposition 4.12, the cylinder in Example 4.10 is orientable,
while the Möbius strip in Example 4.11 is not orientable.
Example 4.15. Recall that the real projective space RP2 consists of homogeneous triples
[x0 : x1 : x2 ] where (x0 , x1 , x2 ) 6= (0, 0, 0). The standard parametrizations are given by:
F0 (x1 , x2 ) = [1 : x1 : x2 ]
F1 (y0 , y2 ) = [y0 : 1 : y2 ]
F2 (z0 , z1 ) = [z0 : z1 : 1]
By the fact that [y0 : 1 : y2 ] = [1 : y0−1 : y2 y0−1 ], the transition map F0−1 ◦ F1 is defined on
{(y0 , y2 ) ∈ R2 : y0 6= 0}, and is given by: (x1 , x2 ) = (y0−1 , y2 y0−1 ). Hence,
−y0−2
∂(x1 , x2 ) 0
D(F0−1 ◦ F1 ) = =
∂(y0 , y2 ) −y2 y0−2 y0−1
1
det D(F0−1 ◦ F1 ) = − 3
y0
Therefore, it is impossible for det D(F0−1 ◦ F1 ) > 0 on the overlap domain {(y0 , y2 ) ∈ R2 :
y0 6= 0}.
At this stage, we have shown that this altas is not an oriented one. In order to prove
RP2 is non-orientable, we need to show any altas of RP2 is not oriented. We will prove
this using Proposition 4.25 later.
122 4. Generalized Stokes’ Theorem
Exercise 4.9. Show that RP3 is orientable. Propose a conjecture about the ori-
entability of RPn .
Exercise 4.10. Show that for any smooth manifold M (whether or not it is ori-
entable), the tangent bundle T M must be orientable.
Exercise 4.11. Show that for a smooth orientable manifold M with boundary, the
boundary manifold ∂M must also be orientable.
4.3. Integrations of Differential Forms 123
From the definition, we see that it only makes sense to integrate an n-form on an
n-dimensional manifold.
Very few manifolds can be covered by a single parametrization. Of course, Rn
is an example. One less trivial example is the graph of a smooth function. Suppose
f (x, y) : R2 → R is a smooth function. Consider its graph:
Γf := {(x, y, f (x, y)) ∈ R3 : (x, y) ∈ R2 }
which can be globally parametrized by F : R2 → Γf where
F (x, y) = (x, y, f (x, y)).
2 2
Let ω = e−x −y
dx ∧ dy be a 2-form on Γf , then its integral over Γf is given by:
Z Z Z ∞Z ∞
−x2 −y 2 2 2
ω= e dx ∧ dy = e−x −y dx dy = π.
Γf Γf −∞ −∞
which is not consistent with the previous result. How shall we fix it?
(2) Even if a manifold can be covered by one single parametrization, such a parametriza-
tion may not be unique. If both (u1 , . . . , un ) and (v1 , . . . , vn ) are global coordinates
of M , then a differential form ω can be expressed in terms of either ui ’s or vi ’s. Is
the integral independent of the chosen coordinate system?
The first issue can be resolved easily. Whenever we talk about integration of differential
forms, we need to first fix the order of the coordinates. Say on R2 we fix the order to be
(x, y), then for any given 2-form we should express it in terms of dx ∧ dy before “erasing
the wedges”. For the 2-form ω above, we must first express it as:
2
−y 2
ω = e−x dx ∧ dy
before integrating it.
124 4. Generalized Stokes’ Theorem
Let’s examine the second issue. Suppose M is an n-manifold with two different
global parametrizations F (u1 , . . . , un ) : U → M and G(v1 , . . . , vn ) : V → M . Given an
n-form ω which can be expressed as:
ω = ϕ du1 ∧ · · · ∧ dun ,
then from Proposition 3.56, ω can be expressed in terms of vi ’s by:
∂(u1 , . . . , un ) 1
ω = ϕ det dv ∧ · · · ∧ dv n .
∂(v1 , . . . , vn )
Recall that the change-of-variable formula in Multivariable Calculus asserts that:
Z Z
1 n ∂(u1 , . . . , un )
ϕ du · · · du = ϕ det dv 1 · · · dv n .
U V ∂(v 1 , . . . , v n )
Z
Therefore, in order for ω to be well-defined, we need
M
Z Z
∂(u1 , . . . , un ) 1
ϕ du1 ∧ · · · ∧ dun and ϕ det dv ∧ · · · ∧ dv n
F (U ) F (V) ∂(v1 , . . . , vn )
to be equal, and so we require:
∂(u1 , . . . , un )
det > 0.
∂(v1 , . . . , vn )
When defining an integral of a differential form, we not only need to choose a
convention on the order of coordinates, say (u1 , . . . , un ), but also we shall only consider
∂(u1 , . . . , un )
those coordinate systems (v1 , . . . , vn ) such that det > 0. Therefore, in
∂(v1 , . . . , vn )
order to integrate a differential form, we require the manifold to be orientable.
Example 4.16. Let S2 be the unit sphere in R3 centered at the origin. Consider the
2-form ω on R3 defined as:
ω = dx ∧ dy.
2 3
Let ι : S → R be the inclusion
Z map, then ι∗ ω is a 2-form on S2 . We are interested in
the value of the integral ι∗ ω.
S2
4.3. Integrations of Differential Forms 125
inclusion map.
Next, we will discuss how to define integrals of differential forms when M is covered
by multiple parametrizations none of which can almost cover the whole manifold. The
key idea is to break down the n-form into small pieces, so that each piece is completely
covered by one single parametrization. It will be done using partition of unity to be
discussed.
We first introduce the notion of support which appears often in the rest of the course
(as well as in advanced PDE courses).
Remark 4.19. It can be shown that given any smooth manifold with any atlas, partitions
of unity subordinate to that given atlas must exist. The proof is very technical and is
not in the same spirit with other parts of the course, so we omit the proof here. It is
more important to know what partitions of unity are for, than to know the proof of
existence.
Remark 4.20. Note that partitions of unity subordinate to a given atlas may not be
unique!
Remark 4.21. Condition
P (ii) in Definition 4.18 is merely a technical analytic condition
to make sure the sum all α ρα (p) is a finite sum for each fixed p ∈ M , so that we do not
need to worry about convergence issues. If the manifold can be covered by finitely many
local parametrizations, then condition (ii) automatically holds (and we do not need to
worry about).
Recall that every open cover of a compact set has a finite sub-cover. Together with
condition (ii) in Definition 4.18, one can show that ρα ω are identically zero for all except
finitely many α’s. The argument goes as follows: at each p ∈ supp ω, by condition (ii) in
Definition 4.18, there exists an open set Op ⊂ M containing p such that the set:
Sp := {α : supp ρα ∩ Op 6= ∅}
Therefore, if α is an index such that supp (ρα ω) 6= ∅, then there exists i ∈ {1, . . . , N }
such that supp ρα ∩ Opi 6= ∅, or in other words, α ∈ Spi for some i, and so:
N
[
{α : supp (ρα ω) 6= ∅} ⊂ Spi .
i=1
Since each Spi is a finite set, the setZ{α : supp (ρα ω) 6= ∅} is also finite. Therefore, there
are only finitely many α’s such that is non-zero, and so the sum stated in (4.5) is
Fα (Uα )
in fact a finite sum.
Now we have understood that there is no convergence issue for (4.5) provided that
ω has compact support (which is automatically true if the manifold M is itself compact).
There are still two well-definedness issues to resolve, namely whether the integral in (4.5)
is independent of oriented atlas A, and for each atlas whether the integral is independent
of the choice of partitions of unity.
128 4. Generalized Stokes’ Theorem
Proposition 4.22. Let M n be an orientable smooth manifold with two oriented atlas
A = {Fα : Uα → M } and B = {Gβ : Vβ → M }
such that det D(Fα−1
◦ Gβ ) > 0 on the overlap for any pair of α and β. Suppose {ρα :
M → [0, 1]} and {σβ : M → [0, 1]} are partitions of unity subordinate to A and B
respectively. Then, given any compactly supported differential n-form ω on M n , we have:
XZ XZ
ρα ω = σβ ω.
all α Fα (Uα ) all β Gβ (Vβ )
X
Proof. By the fact that σβ ≡ 1 on M , we have:
all β
XZ XZ X XXZ
ρα ω = σβ ρα ω =
ρα σβ ω.
all α Fα (Uα ) all α Fα (Uα ) all β all α all β Fα (Uα )∩Gβ (Vβ )
The last equality follows from the fact that supp σβ ⊂ Gβ (Vβ ).
One can similarly work out that
XZ XXZ
σβ ω = ρα σβ ω.
all β Gβ (Vβ ) all β all α Fα (Uα )∩Gβ (Vβ )
P P
Note that α β is a finite double sum and so there is no issue of switching them. It
completes the proof.
By Proposition 4.22, we justified that (4.5) is independent of oriented atlas and the
choice of partitions of unity. We can now define:
We claim ω(p) 6= 0 at every point p ∈ M . Suppose p ∈ Fβ (Uβ ) for some β in the atlas. By
(3.19), for each α, locally near p we have:
∂(u1α , . . . , unα ) 1
du1α ∧ · · · ∧ dunα = det du ∧ · · · ∧ dunβ ,
∂(u1β , . . . , unβ ) β
and so: !
X ∂(u1α , . . . , unα )
ω= ρα det du1β ∧ · · · ∧ dunβ .
∂(u1β , . . . , unβ )
all α
X ∂(u1α , . . . , unα )
Since ρα ≥ 0, ρα ≡ 1 and det > 0, we must have:
∂(u1β , . . . , unβ )
all α
X ∂(u1α , . . . , unα )
ρα det >0 near p.
∂(u1β , . . . , unβ )
all α
Recall that when we integrate an n-form, we need to first pick an order of local
coordinates (u1 , . . . , un ), then express the n-form according to this order, and locally
define the integral as:
Z Z
ϕ du1 ∧ · · · ∧ dun = ϕ du1 · · · dun .
F (U ) U
Note that picking the order of coordinates is a local notion. To rephrase it using global
terms, we can first pick an orientation Ω (which is a global object on M ), then we
require the order of any local coordinates (u1 , . . . , un ) to be Ω-oriented. Any pair of
local coordinate systems (u1 , . . . , un ) and (v1 , . . . , vn ) which are both Ω-oriented will
∂(u1 , . . . , un )
automatically satisfy det > 0 on the overlap.
∂(v1 , . . . , vn )
To summarize, given an orientable manifold M n with a chosen orientation Ω, then
for any local coordinate system F (u1 , . . . , un ) : U → M , we define:
(R
ϕ du1 · · · dun if (u1 , . . . , un ) is Ω-oriented
Z
ϕ du1 ∧ · · · ∧ dun = UR
1 n
F (U ) − U ϕ du · · · du if (u1 , . . . , un ) is not Ω-oriented
or to put it in a more elegant (yet equivalent) way:
Z Z
∂ ∂
ϕ du1 ∧ · · · ∧ dun = sgn Ω , ..., ϕ du1 · · · dun .
F (U ) ∂u 1 ∂u n U
Exercise 4.14. Let Ω := dx∧dy ∧dz be the orientation of R3 . Which of the following
is Ω-oriented?
(a) local coordiantes (x, y, z)
(b) vectors {î, k̂, ĵ}
(c) vectors {u, v, u × v} where u and v are linearly independent vectors in R3 .
Exercise 4.15. Consider three linearly independent vectors {u, v, w} in R3 such that
u ⊥ w and v ⊥ w. Show that {u, v, w} has the same orientation as {î, ĵ, k̂} if and
only if w = cu × v for some positive constant c.
Proposition 4.25 can be used to complete the proof that RP2 is not orientable in Ex-
ample 4.15. In that example, we demonstrated that there are two local parametrizations
F0 (u1 , u2 ) and F1 (v1 , v2 ) with the properties that:
• the domain of each of Fi is connected; while
• their overlap, i.e. domain of F0−1 ◦ F1 , is not connected; and
• det D(F0−1 ◦ F1 ) is positive on one component U , but negative on another compo-
nent V .
4.3. Integrations of Differential Forms 131
To show that RP2 is not orientable, we argue by contradiction that there exists a global
∂
non-vanishing 2-form Ω. Then, if Ω( ∂u , ∂ ) > 0, then one has Ω( ∂v∂ 1 , ∂v∂ 2 ) > 0 on U
1 ∂u2
since det D(F0 ◦ F1 ) > 0 on U , and Ω( ∂v∂ 1 , ∂v∂ 2 ) < 0 on V since det D(F0−1 ◦ F1 ) < 0.
−1
However, since the domain of F1 (v1 , v2 ) is connected and Ω( ∂v∂ 1 , ∂v∂ 2 ) is a smooth (in
particular continuous) function on that domain, there must be a point p in the domain
of F1 such that Ω( ∂v∂ 1 , ∂v∂ 2 ) = 0 at p. It leads to a contradiction that Ω is non-vanishing.
∂
Similar for the case Ω( ∂u , ∂ ) < 0.
1 ∂u2
132 4. Generalized Stokes’ Theorem
The “extra” factor of (−1)n does not look nice at the first glance, but as we will
see later, it will make Generalized Stokes’ Theorem nicer. We are now ready to state
Generalized Stokes’ Theorem in a precise way:
4.4. Generalized Stokes’ Theorem 133
4.4.2. Proof of Generalized Stokes’ Theorem. The proof consists of three steps:
Step 1: a special case where supp ω is contained inside a single parametrization chart
of interior type;
Step 2: another special case where supp ω is contained inside a single parametrization
chart of boundary type;
Step 3: use partitions of unity to deduce the general case.
Proof of Theorem 4.27. Throughout the proof, we will let Ω be the orientation of M ,
and iη Ω be the orientation of ∂M with η being an outward-point normal vector to ∂M .
All local coordinate system (u1 , . . . , un ) of M is assumed to be Ω-oriented.
Step 1: Suppose supp ω is contained in a single parametrization chart of interior type.
Let F (u1 , . . . , un ) : U ⊂ Rn → M be a local parametrization of interior type such
that supp ω ⊂ F (U). Denote:
du1 ∧ · · · ∧ du
ci ∧ · · · ∧ dun := du1 ∧ · · · ∧ dui−1 ∧ dui+1 ∧ · · · ∧ dun ,
Since supp ω ⊂ F (U), the functions ωi ’s are identically zero near and outside the bound-
ary of U ⊂ Rn . Therefore, we can replace the domain of integration U of the RHS integral
134 4. Generalized Stokes’ Theorem
X n Z R Z R
u =R
= (−1)i−1 ··· [ωi ]uii =−R du1 · · · du
ci · · · dun .
i=1 −R −R
n
Since ωi ’s vanish at the boundary of
Z the rectangle [−R, R] , we have ωi = 0 when
ui = ±R. As a result, we proved dω = 0. Since supp ω is contained in a single
M
parametrization
Z chart of interior type, we have ω = 0 on the boundary ∂M . Evidently,
we have ω = 0 in this case. Hence, we proved
∂M
Z Z
dω = ω=0
M ∂M
in this case.
Step 2: Suppose supp ω is contained inside a single parametrization chart of boundary type.
Let G(u1 , . . . , un ) : V ⊂ Rn+ → M be a local parametrization of boundary type such
that supp ω ⊂ G(V). As in Step 1, we express
n
X
ω= ωi du1 ∧ · · · ∧ du
ci ∧ · · · ∧ dun .
i=1
Now V is an open set in Rn+ instead of Rn . Recall that the boundary is the set of
points with un = 0. Therefore, this time we replace V by the half-space rectangle
[−R, R] × · · · × [−R, R] × [0, R] where R > 0 again is a sufficiently large number.
One key difference from Step 1 is that even though ωi ’s has compact support in-
side V, it may not vanish on the boundary of M . Therefore, we can only guarantee
ωi (u1 , . . . , un ) = 0 when un = R, but we cannot claim ωi = 0 when un = 0. Some more
work needs to be done:
Z Z X n
∂ωi 1
dω = (−1)i−1 du · · · dun
M V i=1 ∂ui
Z RZ R Z R X n
∂ωi 1
= ··· (−1)i−1 du · · · dun
0 −R −R i=1 ∂ui
n−1 Z RZ R Z R
X ∂ωi 1
= (−1)i−1 ··· (−1)i−1 du · · · dun
i=1 0 −R −R ∂u i
Z RZ R Z R
∂ωn 1
+ (−1)n−1 ··· du · · · dun
0 −R −R ∂un
4.4. Generalized Stokes’ Theorem 135
!
X X
ω= ρα ω= ρα ω
α α
| {z }
≡1
Z Z X XZ
ω= ρα ω = ρα ω.
∂M ∂M α α ∂M
XZ XZ
ρα ω = d(ρα ω)
α ∂M α M
XZ
= (dρα ∧ ω + ρα dω)
α M
Z ! !
X X
= d ρα ∧ω+ ρα dω.
M α α
!
X X
Since ρα ≡ 1 and hence d ρα ≡ 0, we have proved:
α α
Z XZ Z Z
ω= ρα ω = 0 ∧ ω + 1 dω = dω.
∂M α ∂M M M
Remark 4.28. As we can see from that the proof (Step 2), if we simply choose an
orientation for ∂M such that (u1 , . . . , un−1 ) becomes the order of local coordinates for
∂M , then (4.7) would have a factor of (−1)n on the RHS, which does not look nice.
Moreover, if we pick i−η Ω to be the orientation of ∂M (here −η is then an inward-
pointing normal to ∂M ), then the RHS of (4.7) would have a minus sign, which is not
nice either.
Corollary 4.29 (Green’s Theorem). Let R be a closed and bounded smooth 2-submanifold
in R2 with boundary ∂R. Given any smooth vector field V = (P (x, y), Q(x, y)) defined in
R, then we have: I Z
∂Q ∂P
V · dl = − dx dy,
∂R R ∂x ∂y
∂ ∂
The line integral on the LHS is oriented such that { ∂x , ∂y } has the same orientation as
{η, T } where η is the outward-pointing normal of R, and T is the velocity vector of the
curve ∂R. See Figure 4.3.
The only thing left to figure out is the orientation of the line integral. Locally parametrize
R by local coordinates (s, t) so that {t = 0} is the boundary ∂R and {t > 0} is the interior
of R (see Figure 4.3). By convention, the local coordinate s for ∂R must be chosen so
∂
that Ω(η, ∂s ) > 0 where η is a outward-pointing normal vector to ∂R. In other words,
∂ ∂ ∂
the pair {η, ∂s } should have the same orientation as { ∂x , ∂y }. According to Figure 4.3,
we must choose the local coordinate s for ∂R such that for the outer boundary, s goes
counter-clockwisely as it increases; whereas for each inner boundary, s goes clockwisely
as it increases.
However, not every surface can be covered (or almost covered) by a single parametriza-
tion chart. Generally, if A = {Fα (uα , vα ) : Uα → R3 } is an oriented atlas of Σ with a
partition of unity {ρα : Σ → [0, 1]} subordinate to A, we then define:
X ∂Fα ∂Fα
dS := ρα × duα dvα .
α
∂uα ∂vα
Corollary 4.30 (Stokes’ Theorem). Let Σ be a closed and bounded smooth 2-submanifold
in R3 with boundary ∂Σ, and V = (P (x, y, z), Q(x, y, z), R(x, y, z)) be a vector field
which is smooth on Σ, then we have:
I Z
V · dl = (∇ × V ) · ν dS.
∂Σ Σ
Here {î, ĵ, k̂} has the same orientation as {η, T, ν}, where η is the outward-point normal
vector of Σ at points of ∂Σ, T is the velocity vector of ∂Σ, and ν is the unit normal vector
to Σ in R3 . See Figure 4.4.
Proof. Define:
ω = P dx + Q dy + R dz
which is viewed as a 1-form on Σ. Then,
I I
(4.8) ω= V · dl.
∂Σ ∂Σ
On each local coordinate chart Fα (Uα ), a normal vector to Σ in R3 can be found using
cross products:
∂Fα ∂Fα ∂(y, z) ∂(z, x) ∂(x, y)
× = det î + det ĵ + det k̂
∂uα ∂vα ∂(uα , vα ) ∂(uα , vα ) ∂(uα , vα )
∂Q ∂P ∂P ∂R ∂R ∂Q
∇×V = − î + − ĵ + − k̂.
∂x ∂y ∂z ∂x ∂y ∂z
Hence,
X ∂Fα ∂Fα
dω = (∇ × V ) · × ρα duα ∧ dvα ,
α
∂uα ∂vα
and so Z
XZ ∂Fα ∂Fα
dω = (∇ × V ) · × ρα duα dvα .
Σ α Uα ∂uα ∂vα
∂Fα ∂Fα
∂uα × ∂vα
X ∂Fα ∂Fα
Denote ν = , and recall the fact that dS := ρα × duα dvα ,
∂F
× ∂F
α
∂uα ∂vα
∂uα ∂vα
we get:
Z Z
(4.9) dω = (∇ × V ) · ν dS.
Σ Σ
Combining the results of (4.8) and (4.9), using Generalized Stokes’ Theorem (Theorem
4.7, we get: I Z
V · dl = (∇ × V ) · ν dS
∂Σ Σ
as desired.
To see the orientation of ∂Σ, we locally parametrize Σ by coordinates (s, t) such that
{t = 0} are points on ∂Σ, and so ∂Σ is locally parametrized by s. The outward-pointing
∂
∂
normal of is given by η := − ∂t
n ∂Σ in Σ o . By convention, the orientation of η, ∂s is the
∂ ∂
same as ∂uα , ∂vα , and hence:
∂ ∂ ∂
η, , ν has the same orientation as , ,ν .
∂s ∂uα ∂vα
∂Fα ∂Fα
∂uα × ∂vα
n
∂ ∂
o
As ν = , the set ∂uα , ∂vα , ν has the same orientation as {î, ĵ, k̂}. As
∂Fα ∂Fα
∂uα × ∂vα
∂
a result, the set {η, ∂s , ν} is oriented in the way as in Figure 4.4.
Using {î, ĵ, k̂} as the orientation for D, then it is clear that:
Z Z
(4.10) dω = ∇ · V dx dy dz.
D D
Consider an atlas A = {Fα (uα , vα , wα ) : Uα → R3 } of D such that for the local
parametrization of boundary type, the boundary points are given by {wα = 0}, and
interior points are {wα > 0}. Then, ∂D is locally parametrized by (uα , vα ).
∂
As a convention, the orientation of (uα , vα ) is chosen such that {− ∂w , ∂ , ∂ }
α ∂uα ∂vα
has the same orientation as {î, ĵ, k̂}, or equivalently, { ∂u∂α , ∂v∂α , − ∂w
∂
α
} has the same
orientation as {î, ĵ, k̂}.
∂Fα ∂Fα
∂uα × ∂vα
Furthermore, let ν be the unit normal of ∂D given by ν = . By the
∂Fα ∂Fα
∂uα × ∂vα
convention of cross products, { ∂F α ∂Fα
∂uα , ∂vα , ν} must have the same orientation as {î, ĵ, k̂}.
∂ ∂ ∂ ∂Fα ∂Fα ∂
Now that { ∂uα , ∂vα , − ∂wα } and { ∂uα , ∂vα , ν} have the same orientation, so ν and − ∂w α
are both pointing in the same direction. In other words, ν is the outward-point normal.
The rest of the proof goes by writing ω in terms of duα ∧ dvα on each local coordinate
chart:
X
ω= ρα ω
α
X ∂(y, z) ∂(z, x) ∂(x, y)
= ρα P det + Q det + R det duα ∧ dvα
α
∂(uα , vα ) ∂(uα , vα ) ∂(uα , vα )
X ∂Fα ∂Fα
= V · × ρα duα ∧ dvα
α
∂uα ∂vα
X ∂Fα ∂Fα
= V · ν ρα × duα ∧ dvα
α
∂uα ∂vα
Therefore, we get:
I I I
X ∂Fα ∂Fα
(4.11) ω= V · ν ρα × duα dvα = V · ν dS.
∂D ∂D α
∂uα ∂vα ∂D
Combining with (4.10), (4.11) and Generalized Stokes’ Theorem, the proof of this
corollary is completed.
Chapter 5
De Rham Cohomology
Michael Atiyah
141
142 5. De Rham Cohomology
but not exact, it is fair to group ω and ω + dη’s together, and count them as one. In
formal mathematical language, equivalence classes are used as we will discuss in detail.
It turns out that the “number” of closed, not exact k-forms on a given M is a related to
the topology of M !
In this chapter, we will learn the basics of de Rham cohomology, which is a beautiful
topic to end the course MATH 4033.
= {closed k-forms on M }
k−1
T M → ∧ T M = {ω ∈ ∧k T ∗ M : ω = dη for some η ∈ ∧k−1 T ∗ M }
∗ k ∗
Im d : ∧
= {exact k-forms on M }
In many occasions, we may simply denote the above kernel and image by ker(d) and
Im (d) whenever the value of k is clear from the context.
By d2 = 0, it is easy to see that:
Im d : ∧k−1 T ∗ M → ∧k T ∗ M ⊂ ker d : ∧k T ∗ M → ∧k+1 T ∗ M .
If all closed k-forms on a certain manifold are exact, then we have Im (d) = ker(d). How
“many” closed k-forms are exact is then measured by how Im (d) is “smaller” than ker(d),
which is precisely measured by the size of the quotient vector space ker(d)/Im (d). We
call this quotient the de Rham cohomology group1.
Definition 5.1 (de Rham Cohomology Group). Let M be a smooth manifold. For any
positive integer k, we define the k-th de Rham cohomology group of M to be the quotient
vector space:
ker d : ∧k T ∗ M → ∧k+1 T ∗ M
k
HdR (M ) := .
Im (d : ∧k−1 T ∗ M → ∧k T ∗ M )
5.1.1. Quotient Vector Spaces. Let’s first review the basics about quotient vector
spaces in Linear Algebra. Given a subspace W of a vector space V , we can define an
equivalence relation ∼ by declaring that v1 ∼ v2 if and only if v1 − v2 ∈ W . For example,
if W is the x-axis and V is the xy-plane, then two vector v1 and v2 are equivalent under
this relation if and only if they have the same ĵ-component.
1A vector space is also a group whose addition is the vector addition. Although it is more appropriate or precise to call the
quotient the “de Rham cohomology space”, we will follow the history to call it a group.
5.1. De Rham Cohomology 143
For each element v ∈ V (the bigger space), one can define an equivalence class:
[v] := {u ∈ V : u ∼ v} = {u ∈ V : u − v ∈ W }
which is the set of all vectors in V that are equivalent to v. For example, if W is the
x-axis and V is R2 , then the class [(2, 3)] is given by:
[(2, 3)] = {(x, 3) : x ∈ R}
which is the horizontal line {y = 3}. Similarly, one can figure out [(1, 3)] = [(2, 3)] =
[(3, 3)] = . . . as well, but [(2, 3)] 6= [(2, 2)], and the latter is the line {y = 2}.
The quotient space V /W is defined to be the set of all equivalence classes, i.e.
V /W := {[v] : v ∈ V }.
2
For example, if V is R and W is the x-axis, then V /W is the set of all horizontal lines in
R2 . For finite dimensional vector spaces, one can show (see Exercise 5.1) that
dim(V /W ) = dim V − dim W,
and so the “size” (precisely, the dimension) of the quotient V /W measures how small
W is when compared to V . In fact, if the bases of V and W are suitably chosen, we can
describe the basis of V /W in a precise way (see Exercise 5.1).
5.1.2. Cohomology Classes and Betti numbers. Recall that the k-th de Rham
k
cohomology group HdR (M ), where k ≥ 1, of a smooth manifold M is defined to be the
quotient vector space:
ker d : ∧k T ∗ M → ∧k+1 T ∗ M
k
HdR (M ) := .
Im (d : ∧k−1 T ∗ M → ∧k T ∗ M )
Given a closed k-form ω, we then define its equivalence class to be:
[ω] := {ω 0 : ω 0 − ω is exact}
= {ω 0 : ω 0 = ω + dη for some η ∈ ∧k−1 T ∗ M }
= {ω + dη : η ∈ ∧k−1 T ∗ M }.
144 5. De Rham Cohomology
An equivalence class [ω] is called the de Rham cohomology class represented by (or
containing) ω, and ω is said to be a representative of this de Rham cohomology class.
By Exercise 5.1, its dimension is given by
k
dim HdR (M )
= dim ker d : ∧k T ∗ M → ∧k+1 T ∗ M − dim Im d : ∧k−1 T ∗ M → ∧k T ∗ M
Definition 5.3 (Betti Numbers). Let M be a smooth manifold. The k-th Betti number
of M is defined to be:
k
bk (M ) := dim HdR (M ).
0
In particular, b0 (M ) = dim HdR (M ) is the number of connected components of M .
In case when M = R2 \{(0, 0)}, we discussed that there is a closed 1-form
−y dx + x dy
ω=
x2 + y 2
defined on M which is not exact. Therefore, ω ∈ ker d : ∧1 T ∗ M → ∧2 T ∗ M yet ω 6∈
Im d : ∧0 T ∗ M → ∧1 T ∗ M , and so in HdR 1
(M ) we have [ω] 6= [0]. From here we can
1
conclude that HdR (M ) 6= {[0]} and b1 (M ) ≥ 1. We will later show that in fact b1 (M ) = 1
using some tools in later sections.
Exercise 5.3. If k > dim M , what can you say about bk (M )?
1
Theorem 5.4 (Poincaré Lemma for HdR ). For any star-shaped open set U in Rn , we have
1
HdR (U ) = {[0]}. In other words, any closed 1-form defined on a star-shaped open set is
exact on that open set.
where Lx is the line segment joining p and x, which can be parametrized by:
γ(t) = (1 − t)p + tx, t ∈ [0, 1].
Write p = (p1 , . . . , pn ), x = (x1 , . . . , xn ), then f (x) can be expressed in terms of t by:
Z 1X n
f (x) = ωi (γ(t)) · (xi − pi ) dt.
0 i=1
5.1. De Rham Cohomology 145
n Z n
!
1
X X ∂ωi
= t δjk · (xi − pi ) + ωj (γ(t)) dt
i=1 0
∂xk
k=1
n Z 1
X ∂ωi
= t · (xi − pi ) + ωj (γ(t)) dt
i=1 0
∂xj
∂ωi ∂ωj
and hence = for any i, j. Using this to proceed our calculation:
∂xj ∂xi
Z 1
∂f ∂ωj
(x) = t · (xi − pi ) + ωj (γ(t)) dt
∂xj 0 ∂xi
Z 1
d
= (tωj (γ(t))) dt
0 dt
t=1
= [tωj (γ(t))]t=0 = ωj (γ(1)) = ωj (x).
In the second equality above, we have used the chain rule backward:
d ∂ωj
(tωj (γ(t))) = t · (xi − pi ) + ωj (γ(t)).
dt ∂xi
1
From this, we conclude that ω = df on U , and hence [ω] = [0] in HdR (U ). Since ω is
1
an arbitrary closed 1-form on U , we have HdR (U ) = {[0]}.
k
Remark 5.5. Poincaré Lemma also holds for HdR , meaning that if U is a star-shaped
n k
open set in R , then HdR (U ) = {[0]} for any k ≥ 1. However, the proof involves the use
of Lie derivatives and a formula by Cartan, both of which are beyond the scope of this
0
course. Note also that HdR (U ) ' R since a star-shaped open set must be connected.
Remark 5.6. We have discussed that the 1-form
−y dx + x dy
ω=
x2 + y 2
is closed but not exact. To be precise, it is not exact on R2 \{(0, 0)}. However, if we
regard the domain to be the first quadrant U := {(x, y) : x > 0 and y > 0}, which is a
star-shaped open set in R2 , then by Poinaré Lemma (Theorem 5.4), ω is indeed an exact
1-form on U . In fact, it is not difficult to verify that
y
ω = d tan−1 on U .
x
146 5. De Rham Cohomology
Note that the scalar function tan−1 xy is smoothly defined on U . Whether a form is exact
or not depends on the choice of its domain!
k
Theorem 5.7 (Diffeomorphism Invariance of HdR ). If two smooth manifolds M and N
k k
are diffeomorphic, then HdR (M ) and HdR (N ) are isomorphic for any k ≥ 0.
Proof. Combine the results of the Poincaré Lemma (Theorem 5.4) and the diffeomor-
1
phism invariance of HdR (Theorem 5.7).
Under the assumption of Riemann Mapping Theorem (whose proof can be found in
1
advanced Complex Analysis textbooks), we can establish that HdR (U ) = {[0]} for any
2
(non-empty) simply-connected subset U in R . Consequently, any closed 1-form on such
a domain U is exact on U . Using the language in Multivariable Calculus (or Physics),
this means any curl-zero vector field defined on a (non-empty) simply-connected domain
U in R2 must be conservative on U . You might have learned this fact without proof in
MATH 2023.
148 5. De Rham Cohomology
One good way to think of a deformation retract is to regard t as the time, and Ψt
is a “movie” that demonstates how M collapses onto Σ. The condition Ψ0 = idM says
initially (at t = 0), the “movie” starts with the image M . At the final scene (at t = 1), the
condition Ψ1 : M → Σ says that the image eventually becomes Σ . The last condition
Ψt (p) = p for any p ∈ Σ means the points on Σ do not move throughout the movie.
Before we talk about the relation between cohomology and deformation retract, let’s first
look at some examples:
Example 5.10. The unit circle S1 defined by {(x, y) : x2 + y 2 = 1} is a deformation
retract of the annulus {(x, y) : 14 < x2 + y 2 < 4}. To describe such a retract, it’s best to
use polar coordinates:
Ψt (reiθ ) = (r + t(1 − r)) eiθ
For each t ∈ [0, 1], the map Ψt has image inside the annulus since r + t(1 − r) ∈ ( 21 , 2)
whenever r ∈ ( 12 , 2) and t ∈ [0, 1]. One can easily check that Ψ0 (reiθ ) = reiθ , Ψ1 (reiθ ) =
eiθ and Ψt (eiθ ) = eiθ for any (r, θ) and t ∈ [0, 1]. Hence Ψt fulfills all three conditions
stated in Definition 5.9.
Example 5.11. Intuitively, we can see the letters E, F, H, K, L, M and N all deformation
retract onto the letter I. Also, the letter Q deformation retracts onto the letter O. The
explicit Ψt for each deformation retract is not easy to write down.
Example 5.12. A two-dimensional torus with a point removed can deformation retract
onto two circles joined at one point. Try to visualize it!
Exercise 5.5. Show that any star-shaped open set U in Rn deformation retracts
onto its base point.
5.2. Deformation Retracts 149
Exercise 5.6. Let M be a smooth manifold, and Σ0 be the zero section of the tangent
bundle, i.e. Σ0 consists of all pairs (p, 0p ) in T M where p ∈ M and 0p is the zero
vector in Tp M . Show that the zero section Σ0 is a deformation retract of the tangent
bundle T M .
We next show an important result in de Rham theory, which asserts that deformation
retracts preserve the first de Rham cohomology group.
Proof. Let ι : Σ → M be the inclusion map, and {Ψt : M → M }t∈[0,1] be the family
of maps satisfying all conditions stated in Definition 5.9. Then, the pull-back map
ι∗ : ∧1 T ∗ M → ∧1 T ∗ Σ induces a map ι∗ : HdR
1 1
(M ) → HdR (Σ). Also, the map Ψ1 : M → Σ
∗ 1 1
induces a pull-back map Ψ1 : HdR (Σ) → HdR (M ). The key idea of the proof is to show
that Ψ∗1 and ι∗ are inverses of each other as maps between HdR 1 1
(M ) and HdR (Σ).
Let ω be an arbitrary closed 1-form defined on M . Similar to the proof of Poincaré
Lemma (Theorem 5.4), we consider the scalar function f : M → R defined by:
Z
f (x) = ω
Ψt (x)
Here, Ψt (x) is regarded as a curve with parameter t joining Ψ0 (x) = x and Ψ1 (x) ∈ Σ.
We will show the following result:
(5.1) Ψ∗1 ι∗ ω − ω = df
which will imply [ω] = Ψ∗1 ι∗ [ω], or in other words, Ψ∗t ◦ ι∗ = id on HdR
1
(M ).
To prove (5.1),
P we use local coordinates (u1 , . . . , un ), and express ω in terms of local
coordinates ω = i ωi dui . For simplicity, let’s assume that such a local coordinate chart
can cover the whole curve Ψt (x) for t ∈ [0, 1]. We will fix this issue later. For each
t ∈ [0, 1], we write Ψit (x) to be the ui -coordinate of Ψt (x), i.e. Ψit = ui ◦ Ψt . Then, one
can calculate df using local coordinates. The calculation is similar to the one we did in
the proof of Poincaré Lemma (Theorem 5.4):
1
∂Ψit
Z Z X
f (x) = ω= ωi (Ψt (x)) dt
Ψt (x) 0 i
∂t
( ! )
X ∂f X Z 1 ∂ X ∂Ψit
j
(df )(x) = du = ωi (Ψt (x)) dt duj
j
∂uj j 0 ∂u j i
∂t
X Z 1 X ∂ωi ∂Ψ k
∂Ψ i X ∂
i
∂Ψ
t t t
= + ωi (Ψt (x)) dt duj
j
0 ∂uk Ψt (x) ∂uj ∂t i
∂t ∂uj
i,k
150 5. De Rham Cohomology
∂ωi ∂ωk
Next, recall that ω is a closed 1-form, so we have = for any i, k. Using this on
∂uk ∂ui
the first term, and by switching indices of the second term in the integrand, we get:
X Z 1 X ∂ωk k
∂Ψt ∂Ψt i X k
∂ ∂Ψt j
(df )(x) = + ωk (Ψt (x)) dt du
j
0 ∂ui Ψt (x) ∂uj ∂t ∂t ∂uj
i,k k
(Z
1
! ) t=1
X ∂ X ∂Ψkt j
X ∂Ψkt
= ωk (Ψt (x)) dt du = ωk (Ψt (x)) duj
j 0 ∂t ∂uj ∂uj t=0
k j,k
where the last equality follows from the (backward) chain rule.
Denote ιt : Ψt (M ) → M the inclusion map at time t, then one can check that
X
Ψ∗t ι∗t ω(x) = (ιt ◦ Ψt )∗ ω(x) = (ιt ◦ Ψt )∗ ωk duk
k
X
= ωk (ιt ◦ Ψt (x)) d(uk ◦ ιt ◦ Ψt (x))
k
X
= ωk (ιt ◦ Ψt (x)) dΨkt
k
X ∂Ψkt
= ωk (Ψt (x)) duj .
∂uj
j,k
Therefore, we get:
t=1
X ∂Ψkt t=1
df = ωk (Ψt (x)) duj = [Ψ∗t ι∗t ω]t=0 = Ψ∗1 ι∗1 ω − Ψ∗0 ι∗0 ω.
∂uj t=0
j,k
Since Ψ0 = idM and ι0 = idM , we have proved (5.1). In case Ψt (x) cannot be covered by
one single local coordinate chart, one can then modify the above proof a bit by covering
the curve Ψt (x) by finitely many local coordinate charts. It can be done because Ψt (x) is
compact. Suppose 0 = t0 < t1 < . . . < tN = 1 is a partition of [0, 1] such that for each
α, the curve Ψt (x) restricted to t ∈ [tα−1 , tα ] can be covered by a single local coordinate
chart, then we have:
N Z tα
X X ∂Ψit
f (x) = ωi (Ψt (x)) dt.
α=1 tα−1 i
∂t
(5.2) ι∗ Ψ∗1 η = η.
(u1 , . . . , uk , 0, . . . , 0) ∈ Σ.
5.2. Deformation Retracts 151
Such a local coordinate system always exists near Σ by Immersion Theorem (Theorem
Pk
2.42). Locally, denote η = i=1 ηi dui , then
k
X k
X
(Ψ∗1 η)(x) = Ψ∗1 (ηi (x) dui ) = ηi (Ψ1 (x)) d(ui ◦ Ψ1 )
i=1 i=1
k X
k
X ∂Ψi1 (x) j
= ηi (Ψ1 (x)) du .
i=1 j=1
∂uj
Since Ψ1 (x) = x whenever x ∈ Σ, we have Ψi1 (x) = ui (x) where ui (x) is the i-th
∂Ψi1 (x) ∂ui
coordinate of x. Therefore, we get = = δij and so:
∂uj ∂uj
k
X k
X
(Ψ∗1 η)(x) = ηi (x)δij duj = ηi (x) dui = η(x)
i,j=1 i=1
∗
for any x ∈ Σ. In other words, ι Ψ∗1 η = η on Σ. This proves (5.2).
Combining (5.1) and (5.2), we get ι∗ ◦ Ψ∗1 = id on HdR1
(Σ), and Ψ∗1 ◦ ι∗ = id on
1
HdR (M ). As a result, Ψ∗1 and ι∗ are inverses
of each other in HdR1
. It completes the proof
1 1
that HdR (M ) and HdR (Σ) are isomorphic.
1
Using Theorem 5.13, we see that HdR (R2 \{(0, 0)}) and HdR
1
(S1 ) are isomorphic, and
2 1
hence b1 (R \{(0, 0)}) = b1 (S ). At this moment, we still don’t know the exact value of
b1 (S1 ), but we will figure it out in the next section.
k
Note that Theorem 5.13 holds for HdR for any k ≥ 2 as well, but the proof again uses
some Lie derivatives and Cartan’s formula, which are beyond the scope of this course.
Another nice consequence of Theorem 5.13 is the 2-dimensional case of the following
celebrated theorem in topology:
Theorem 5.14 (Brouwer’s Fixed-Point Theorem on R2 ). Let B1 (0) be the closed ball
with radius 1 centered at origin in R2 . Suppose Φ : B1 (0) → B1 (0) is a smooth map
between B1 (0). Then, there exists a point x ∈ B1 (0) such that Φ(x) = x.
Proof. We prove by contradiction. Suppose Φ(x) 6= x for any x ∈ B1 (0). Then, we let
Ψt (x) be a point in B1 (0) defined in the following way:
(1) Consider the vector x − Φ(x) which is non-zero.
(2) Consider the straight ray emanating from x in the direction of x − Φ(x). This ray
will intersect the unit circle S1 at a unique point px .
(3) We then define Ψt (x) := (1 − t)x + tpx
We leave it as an exercise for readers to write down the explicit formula for Ψt (x), and
show that it is smooth for each t ∈ [0, 1].
Clearly, we have Ψ0 (x) = x for any x ∈ B1 (0); Ψ1 (x) = px ∈ S1 ; and if |x| = 1, then
px = x and so Ψt (x) = x.
Therefore, it shows S1 is a deformation retract of B1 (0), and by Theorem 5.13,
1 1 1
their HdR ’s are isomorphic. However, we know HdR (B1 (0)) ' {[0]}, while HdR (S1 ) '
1
HdR (R2 \{(0, 0)}) 6= {[0]}. It is a contradiction! It completes the proof that there is at
least a point x ∈ B1 (0) such that Φ(x) = x.
Exercise 5.8. Write down an explicit expression of px in the above proof, and hence
show that Ψt is smooth for each fixed t.
152 5. De Rham Cohomology
Exercise 5.9. Generalize the Brouwer’s Fixed-Point Theorem in the following way:
given a manifold Ω which is diffeomorphic to B1 (0), and a smooth map Φ : Ω → Ω.
Using Theorem 5.14, show that there exists a point p ∈ Ω such that Φ(p) = p.
Exercise 5.10. What fact(s) are needed to be established in order to prove the
Brouwer’s Fixed-Point Theorem for general Rn using a similar way as in the proof of
Theorem 5.14?
5.3. Mayer-Vietoris Theorem 153
One can also talk about exact-ness for a finite sequence, say:
1 T 2 3 T T n Tn−1 T
G0 −→ G1 −→ G2 −→ · · · −−−→ Gn−1 −−→ Gn
However, such a T1 would not have a previous map, and such an Tn would not have the
next map. Therefore, whenever we talk about the exact-ness of a finite sequence of maps,
we will add two trivial maps at both ends, i.e.
0 1 T
2 3 T n T T 0
(5.3) 0−
→ G0 −→ G1 −→ G2 −→ · · · Gn−1 −−→ Gn −
→ 0.
0
The first map 0 − → G0 is the homomorphism taking the zero in the trivial group to the
0
zero in G0 . The last map Gn −→ 0 is the linear map that takes every element in Gn to the
zero in the trivial group. We say the finite sequence (5.3) an exact sequence if
0 0
Im (0 −
→ G0 ) = ker T1 , Im Tn = ker(Gn −
→ 0), and Im Ti = ker Ti+1 for any i.
0 0
Note that Im (0 −
→ G0 ) = {0} and ker(Gn −
→ 0) = Gn , so if (5.3) is an exact sequence, it
is necessary that
ker T1 = {0} and Im Tn = Gn
or equivalently, T1 is injective and Tn is surjective.
One classic example of a finite exact sequence is:
ι f
0→Z→
− C−
→ C\{0} → 0
where ι : Z → C is the inclusion map taking n ∈ Z to itself n ∈ C. The map f : C →
C\{0} is the map taking z ∈ C to e2πiz ∈ C\{0}.
It is clear that ι is injective and f is surjective (from Complex Analysis). Also, we have
Im ι = Z and ker f = Z as well (note that the identity of C\{0} is 1, not 0). Therefore,
this is an exact sequence.
U ∩V M
iV jV
V
where iU , iV , jU and jV are inclusion maps. Each inclusion map, say jU : U → M ,
induces a pull-back map jU∗ : ∧k T ∗ M → ∧k T ∗ U which takes any k-form ω on M , to the
k-form ω|U restricted on U , i.e. jU∗ (ω) = ω|U for any ω ∈ ∧k T ∗ M . In terms of local
expressions,P there is essentially no difference
P between ω and ω|U since U is open. If
locally ω = i ωi dui on M , then ω|U = i ωi dui as well. The only difference is the
domain: ω(p) is defined for every p ∈ M , while ω|U (p) is defined only when p ∈ U .
To summarize, we have the following diagram:
U
i∗
U
∗
jU
U ∩V M
∗
i∗
V
jV
V
Using the pull-backs of these four inclusions iU , iV , jU and jV , one can form a
sequence of linear maps for each integer k:
j ∗ ⊕j ∗ i∗ −i∗
(5.4) 0 → ∧k T ∗ M −−
U V
−−→ ∧k T ∗ U ⊕ ∧k T ∗ V −−
U V
−−→ ∧k T ∗ (U ∩ V ) → 0
Here, ∧k T ∗ U ⊕∧k T ∗ V is the direct sum of the vector spaces ∧k T ∗ U and ∧k T ∗ V , meaning
that:
∧k T ∗ U ⊕ ∧k T ∗ V = {(ω, η) : ω ∈ ∧k T ∗ U and η ∈ ∧k T ∗ V }.
The map jU∗ ⊕ jV∗ : ∧k T ∗ M → ∧k T ∗ U ⊕ ∧k T ∗ V is defined by:
(jU∗ ⊕ jV∗ )(ω) = (jU∗ ω, jV∗ ω) = ( ω|U , ω|V ).
i∗ −i∗
The map ∧k T ∗ U ⊕ ∧k T ∗ V −−
U V
−−→ ∧k T ∗ (U ∩ V ) is given by:
(i∗U − i∗V )(ω, η) = i∗U ω − i∗V η = ω|U ∩V − η|U ∩V .
We next show that the sequence (5.4) is exact. Let’s first try to understand the image
and kernel of each map involved.
Given (ω, η) ∈ ker(i∗U − i∗V ), we will have ω|U ∩V = η|U ∩V . Therefore, ker(i∗U − i∗V )
consists of pairs (ω, η) where ω and η agree on the intersection U ∩ V .
Now consider Im (jU∗ ⊕ jV∗ ), which consists of pairs of the form ( ω|U , ω|V ). Certainly,
the restrictions of both ω|U and ω|V on the intersection U ∩ V are the same, and hence
the pair is inside ker(i∗U − i∗V ). Therefore, we have Im (jU∗ ⊕ jV∗ ) ⊂ ker(i∗U − i∗V ).
5.3. Mayer-Vietoris Theorem 155
Proposition 5.15. Let M be a smooth manifold. Suppose there are two open subsets
U and V of M such that M = U ∪ V , and U ∩ V is non-empty, then the sequence of
maps (5.4) is exact.
Proof. So far we have proved that jU∗ ⊕ jV∗ is injective, and Im (jU∗ ⊕ jV∗ ) ⊂ ker(i∗U − i∗V ).
We next claim that ker(i∗U − i∗V ) ⊂ Im (jU∗ ⊕ jV∗ ):
Let (ω, η) ∈ ker(i∗U − i∗V ), meaning that ω is a k-form on U , η is a k-form on V , and
that ω|U ∩V = η|U ∩V . Define a k-form σ on M = U ∪ V by:
(
ω on U
σ=
η on V
Note that σ is well-defined on U ∩ V since ω and η agree on U ∩ V . Then, we have:
(ω, η) = ( σ|U , σ|V ) = (jU∗ σ, jV∗ σ) = (jU∗ ⊕ jV∗ )σ ∈ Im (jU∗ ⊕ jV∗ ).
Since (ω, η) is arbitrary in ker(i∗U − i∗V ), this shows:
ker(i∗U − i∗V ) ⊂ Im (jU∗ ⊕ jV∗ ).
Finally, we show i∗U − i∗V is surjective. Given any k-form θ ∈ ∧k T ∗ (U ∩ V ), we need
to find a k-form ω 0 on U , and a k-form η 0 on V such that ω 0 − η 0 = θ on U ∩ V . Let
{ρU , ρV } be a partition of unity subordinate to {U, V }. We define:
(
ρV θ on U ∩ V
ω0 =
0 on U \V
Note that ω 0 is smooth: If p ∈ supp ρV ⊂ V , then p ∈ V (which is open) and so ω 0 = ρV θ
in an open neighborhood of p. Note that ρV and θ are smooth at p, so ω 0 is also smooth
at p. On the other hand, if p 6∈ supp ρV , then ω 0 = 0 in an open neighborhood of p. In
particular, ω 0 is smooth at p.
Similarly, we define: (
0 −ρU θ on U ∩ V
η =
0 on V \U
which can be shown to be smooth in a similar way.
Then, when restricted to U ∩ V , we get:
ω 0 |U ∩V − η 0 |U ∩V = ρV θ + ρU θ = (ρV + ρU ) θ = θ.
In other words, we have (i∗U − i∗V )(ω 0 , η 0 ) = θ. Since θ is arbitrary, we proved i∗U − i∗V is
surjective.
k
Recall that a pull-back map on k-forms induces a well-defined pull-back map on HdR .
The sequence of maps (5.4) between space of wedge products induces a sequence of
maps between de Rham cohomology groups:
k U Vj ∗ ⊕j ∗
k k U V k i∗ −i∗
(5.5) 0 → HdR (M ) −−−−→ HdR (U ) ⊕ HdR (V ) −−−−→ HdR (U ∩ V ) → 0.
156 5. De Rham Cohomology
However, the sequence (5.5) is not exact because jU∗ ⊕ jV∗ may not be injective, and
i∗U − i∗V may not be surjective. For example, take M = R2 \{(0, 0)}, and define using
polar coordinates the open sets U = {reiθ : r > 0, θ ∈ (0, 2π)} and V = {reiθ : r > 0, θ ∈
1 1
(−π, π)}. Then, both U and V are star-shaped and hence both HdR (U ) and HdR (V ) are
trivial. Nonetheless we have exhibited that HdR (M ) is non-trivial. The map jU ⊕ jV∗ from
1 ∗
Exercise 5.13. Find an example of M , U and V such that the map i∗U − i∗V in (5.5)
is not surjective.
Nonetheless, it is still true that ker(i∗U − i∗V ) = Im (jU∗ ⊕ jV∗ ), and we will verify it in
the proof of Mayer-Vietoris Theorem (Theorem 5.16). Mayer-Vietoris Theorem asserts
that although (5.5) is not exact in general, but we can connect each short sequence
below:
0 U Vj ∗ ⊕j ∗
0 0 U V 0 i∗ −i∗
HdR (M ) −−−−→ HdR (U ) ⊕ HdR (V ) −−−−→ HdR (U ∩ V )
1 U Vj ∗ ⊕j ∗
1 1 U V 1 i∗ −i∗
HdR (M ) −−−−→ HdR (U ) ⊕ HdR (V ) −−−−→ HdR (U ∩ V )
2 U Vj ∗ ⊕j ∗
2 2 U V 2 i∗ −i∗
HdR (M ) −−−−→ HdR (U ) ⊕ HdR (V ) −−−−→ HdR (U ∩ V )
..
.
to produce a long exact sequence.
The proof of Theorem 5.16 is purely algebraic. We will learn the proof after looking
at some examples.
Therefore, δ is surjective.
By First Isomorphism Theorem (Theorem 5.17), we know:
H 0 (U ∩ V )
1
HdR (M ) = Im δ ∼
= dR .
ker δ
0
Elements of HdR (U ∩ V ) are locally constant functions of the form:
(
a on left arc
fa,b =
b on right arc
Since the Mayer-Vietoris sequence is exact, we have ker δ = Im (i∗U − i∗V ). The space
0 0 0
HdR (U ),HdR (V ) and HdR (U ∩ V ) consist of locally constant functions on U , V and U ∩ V
respectively, and the maps i∗U − i∗V takes constant functions (k1 , k2 ) ∈ HdR0 0
(U ) ⊕ HdR (V )
to the constant function fk1 −k2 ,k1 −k2 on U ∩ V . Therefore, the first de Rham cohomology
group of M is given by:
{fa,b : a, b ∈ R} ∼ R2
1
HdR (M ) ∼
= = ,
{fa−b,a−b : a, b ∈ R} {(x, y) : x = y}
1
and hence b1 (M ) = dim HdR (M ) = 2 − 1 = 1.
Example 5.19. Let’s discuss some consequences of the result proved in the previous
example. Recall that R2 \{(0, 0)} deformation retracts to S1 . By Theorem 5.13, we know
1
HdR (R2 \{(0, 0)}) ∼ 1
= HdR (S1 ).
This tells us b1 (R2 \{(0, 0)}) = 1 as well. Recall that the following 1-form:
−y dx + x dy
ω=
x2 + y 2
158 5. De Rham Cohomology
1
is closed but not exact. The class [ω] is then trivial in HdR (R2 \{(0, 0)}). In an one-
dimensional vector space, any non-zero vector spans that space. Therefore, we conclude:
1
HdR (R2 \{(0, 0)} = {c[ω] : c ∈ R}.
where ω is defined as in above.
As a result, if ω 0 is a closed 1-form on R2 \{(0, 0)}, then we must have
[ω 0 ] = c[ω]
for some c ∈ R, and so ω 0 = cω + df for some smooth function f : R2 \{(0, 0)} → R.
Using the language of vector fields, if V (x, y) : R2 \{(0, 0)} → R2 is a smooth
vector field with ∇ × V = 0, then there is a constant c ∈ R and a smooth function
f : R2 \{(0, 0)} → R such that:
!
−y î + xĵ
V=c + ∇f.
x2 + y 2
Exercise 5.15. Show that b1 (S2 ) = 0. Based on this result, show that any curl-zero
vector field defined on R3 \{(0, 0, 0)} must be conservative.
One good technique of using the Mayer-Vietoris sequence (as demonstrated in the
examples and exercises above) is to consider a segment of the sequence that starts and
ends with the trivial space, i.e.
0 → V1 → V2 → · · · → Vn → 0.
If all vector spaces Vi ’s except one of them are known, then the remaining one (at least its
dimension) can be deduced using First Isomorphism Theorem. Below is a useful lemma
which is particularly useful for finding the Betti number of a manifold:
Lemma 5.20. Let the following be an exact sequence of finite dimensional vector spaces:
1T 2 T Tn−1
0 → V1 −→ V2 −→ · · · −−−→ Vn → 0.
Then, we have:
dim V1 − dim V2 + dim V3 − · · · + (−1)n−1 dim Vn = 0
1
In Example 5.18 (about computing HdR (S1 )), the following exact sequence was used:
0
0 → HdR (S1 ) → HdR
0 0
(U ) ⊕ HdR 0
(V ) → HdR 1
(U ∩ V ) → HdR (S1 ) → HdR
1 1
(U ) ⊕ HdR (V )
| {z } | {z } | {z } | {z } | {z }
R R⊕R R⊕R ? 0
1
Using Lemma 5.20, the dimension of HdR (S1 ) can be computed easily:
1
dim R − dim R ⊕ R + dim R ⊕ R − dim HdR (S1 ) = 0
1
which implies dim HdR (S1 ) = 1 (or equivalently, b1 (S1 ) = 1). Although this method does
1
not give a precise description of HdR (S1 ) in terms of inclusion maps, it is no doubt much
easier to adopt.
In the forthcoming examples, we will assume the following facts stated below (which
we have only proved the case k = 1):
k
• HdR (U ) = 0, where k ≥ 1, for any star-shaped region U ⊂ Rn .
k
• If Σ is a deformation retract of M , then HdR ∼ H k (M ) for any k ≥ 1.
(Σ) = dR
5.3.4. Proof of Mayer-Vietoris Theorem. To end this chapter (and this course),
we present the proof of the Mayer-Vietoris’s Theorem (Theorem 5.16). As mentioned
before, the proof is purely algebraic. The key ingredient of the proof applies to many
other kinds of cohomologies as well (de Rham cohomology is only one kind of many
types of cohomology).
160 5. De Rham Cohomology
Euclidean Hypersurfaces
Chapter 6
Geometry of Curves
Paul Rand
One easy way to make sense of a curve is to regard it as the trajectory of a particle.
At any time t, the functions γ 1 (t), . . . , γ n (t) give the coordinates of the particle in Rn .
Assuming all xi (t), where 1 ≤ i ≤ n, are at least twice differentiable, then the first
derivative γ 0 (t) represents the velocity of the particle, its magnitude |γ 0 (t)| is the speed of
the particle, and the second derivative γ 00 (t) represents the acceleration of the particle.
We will mostly study those curves which are infinitely differentiable (i.e. C ∞ ). For
some technical purposes as we will explain later, we only study those curves γ(t) whose
velocity γ 0 (t) is never zero. We call those curves:
165
166 6. Geometry of Curves
Example 6.2. The curve γ(t) = (cos(et ), sin(et )), where t ∈ (−∞, ∞), is a regular curve
since γ 0 (t) = (−et sin(et ), et cos(et )) and |γ 0 (t)| = et 6= 0 for any t.
However, γe(t) = (cos t2 , sin t2 ), where t ∈ (−∞, ∞), is not a regular curve since
0
e0 (0) = 0.
e (t) = (−2t sin t2 , 2t cos t2 ) and so γ
γ
Although both curves γ(t) and γ e(t) represent the unit circle centered at the origin
in R2 , one is regular but another is not. Therefore, the term regular refers to the
parametrization rather than the trajectory.
Theorem 6.3. Given any regular curve γ(t) : I → Rn , one can always reparametrize it by
arc-length. Precisely, let t0 ∈ I be a fixed number and consider the following function of t:
Z t
s(t) := |γ 0 (τ )| dτ.
t0
Then, t can be regarded as a C ∞ function of s, and the reparametrized curve γ(s) := γ(t(s))
d
is a regular curve such that ds γ(s) = 1 for any s.
Exercise 6.1. Determine whether each of the following is a regular curve. If so,
reparametrize the curve by arc-length:
(a) γ(t) = (cos t, sin t, t), t ∈ (−∞, ∞)
(b) γ(t) = (t − sin t, 1 − cos t), t ∈ (−∞, ∞)
Figure 6.1. curvature measures how fast the unit tangents move
Example 6.6. The circle of radius R centered at the origin (0, 0) on the xy-plane can be
parametrized by γ(t) = (R cos t, R sin t). It can be easily verified that |γ 0 (t)| = R and so
γ(t) is not an arc-length parametrization.
To find an arc-length parametrization, we let:
Z t Z t
0
s(t) = |γ (τ )| dτ = R dτ = Rt.
0 0
s
Therefore, t(s) = Ras a function of s and so an arc-length parametrization of the circle
is: s s
γ(s) := γ(t(s)) = R cos , R sin .
R R
To find its curvature, we compute:
d s s
γ 0 (s) = R cos , R sin
ds s
R
s
R
= − sin , cos
R R
00 1 s 1 s
γ (s) = − cos , − sin
R R R R
1
κ(s) = |γ 00 (s)| = .
R
Thus the curvature of the circle is given by R1 , i.e. the larger the circle, the smaller the
curvature.
Exercise 6.3. Prove that a regular curve γ(t) is a straight line if and only if its
curvature κ is identically zero.
where a and b are positive constants with a 6= b. The arc-length function is given by:
Z tp
s(t) = a2 sin2 τ + b2 cos2 τ dτ.
0
While it is very easy to compute the integral when a = b, there is no closed form or
explicit anti-derivative for the integrand if a 6= b. Although the arc-length parametrization
exists theoretically speaking (Theorem 6.3), it cannot be written down explicitly and so
the curvature cannot be computed from the definition.
The purpose of this section is to derive a formula for computing curvature without
the need of finding its arc-length parametrization. To begin, we first prove the following
important observation:
Lemma 6.7. Let γ(s) : I → Rn be a curve parametrized by arc-length, then the velocity
γ 0 (s) and the acceleration γ 00 (s) is always orthogonal for any s ∈ I.
Proof. Since γ(s) is parametrized by arc-length, we have |γ 0 (s)| = 1 for any s, and so:
d 0 2 d
|γ (s)| = 1=0
ds ds
d 0
(γ (s) · γ 0 (s)) = 0
ds
γ 00 (s) · γ 0 (s) + γ 0 (s) · γ 00 (s) = 0
2γ 00 (s) · γ 0 (s) = 0
γ 00 (s) · γ 0 (s) = 0
Proposition 6.8. Given any regular curve γ(t) in R3 , the curvature as a function of t can
be computed by the following formula:
|γ 0 (t) × γ 00 (t)|
κ(t) = 3 .
|γ 0 (t)|
Proof. Since γ(t) is a regular curve, there exists an arc-length parametrization γ(t(s)),
which for simplicity we denote it by γ(s). From now on, we denote γ 0 (t) as dγ(t) dt ,
0 dγ(s)
regarding t as the parameter of the curve, and γ (s) as ds regarding s as the parameter
of the curve.
170 6. Geometry of Curves
|γ 0 (s)| |γ 00 (s)| = κ(s). Taking the magnitude on both sides of (6.4), we get:
3
dγ d2 γ ds
× 2 =κ .
dt dt dt
Therefore, we get:
|γ 0 (t) × γ 00 (t)|
κ= .
ds 3
dt
The proof can be easily completed by the definition of s(t) and the Fundamental Theorem
of Calculus:
Z t
s= |γ 0 (τ )| dτ
0
ds
= |γ 0 (t)|
dt
Remark 6.9. Since the cross product is involved, Proposition 6.8 can only be used for
curves in R2 or R3 . To apply the result for curves in R2 , say γ(t) = (x(t), y(t)), one may
regard it as the curve γ(t) = (x(t), y(t), 0) in R3 .
By Proposition 6.8, the curvature of the ellipse can be computed easily. See the
example below:
Example 6.10. Let γ(t) = (a cos t, b sin t, 0) be a parametrization of an ellipse on the
xy-plane where a and b are positive constants, then we have:
γ 0 (t) = (−a sin t, b cos t, 0)
γ 00 (t) = (−a cos t, −b sin t, 0)
γ 0 (t) × γ 00 (t) = (ab sin2 t + ab cos2 t) k̂ = ab k̂
Therefore, by Proposition 6.8, it’s curvature function is given by:
|γ 0 (t) × γ 00 (t)| ab
κ(t) = 3 = 2 2 .
|γ 0 (t)| (a sin t + b2 cos2 t)3/2
6.1. Curvature and Torsion 171
Exercise 6.4. Consider the graph of a smooth function y = f (x). Regarding the
graph as a curve in R3 , it can be parametrized using x as the parameter by γ(x) =
(x, f (x), 0). Show that the curvature of the graph is given by:
|f 00 (x)|
κ(x) = 3/2
.
(1 + f 0 (x)2 )
Exercise 6.5. For each of the following curves: (i) compute the curvature κ(t) using
Proposition 6.8; (ii) If it is easy to find an explicit arc-length parametrization of
the curve, compute also the curvature from the definition; (iii) find the (x, y, z)-
coordinates of the point(s) on the curve at which the curvature is the maximum.
(a) γ(t) = (3 cos t, 4 cos t, 5t).
(b) γ(t) = (t2 , 0, t).
(c) γ(t) = 2t, t2 , − 13 t3 .
6.1.5. Frenet-Serret Frame. For now on, we will concentrate on regular curves in
R3 . Furthermore, we consider mostly those curves whose curvature function κ is nowhere
vanishing. Therefore, straight-lines in R3 , or paths such as the graph of y = x3 , are
excluded in our discussion.
We now introduce an important basis of R3 in the studies of space curves, the Frenet-
Serret Frame, or the TNB-frame. It is an orthonormal basis of R3 associated to each point
of a regular curve in R3 .
Remark 6.13. Note that T is a unit vector since γ(s) is arc-length parametrized. Recall
that κ(s) := |γ 00 (s)| and the curve γ(s) is assumed to be non-degenerate. Therefore, N
is well-defined for any s ∈ I and is a unit vector by its definition. From Lemma 6.7, T
and N are orthogonal to each other for any s ∈ I. Therefore, by the definition of cross
product, B is also a unit vector and is orthogonal to both T and N. To conclude, for each
fixed s ∈ I, the Frenet-Serret Frame is an orthonormal basis of R3 .
Example 6.14. Let γ(s) = cos √s2 , sin √s2 , √s2 where s ∈ R. It can be verified easily
that it is arc-length parametrized, i.e. |γ 0 (s)| = 1 for any s ∈ R. The Frenet-Serret Frame
172 6. Geometry of Curves
Remark 6.16. By the definition of the Frenet-Serret Frame, the binormal vector B(s) is
a unit normal vector to the osculating plane Π(s).
Exercise 6.6. Consider the curve γ(t) = (a cos t, a sin t, bt) where a and b positive
constants. First find its arc-length parametrization γ(s) := γ(t(s)), and then compute
its Frenet-Serret Frame.
6.1. Curvature and Torsion 173
Exercise 6.8. [dC76, P.23] Let γ(s) be an arc-length parametrized curve in R3 . The
normal line at γ(s) is the infinite straight line parallel to N(s) passing through the
point represented by γ(s). Suppose the normal line at every γ(s) pass through a
fixed point p ∈ R3 . Show that γ(s) is a part of a circle.
Proof. First note that {T(s), N(s), B(s)} is an orthonormal basis of R3 for any s ∈ I.
Hence, we have:
dB(s)
= a(s)T(s) + b(s)N(s) + c(s)B(s)
ds
where a(s) = dB(s)
ds · T(s), b(s) =
dB(s)
ds · N(s) and c(s) = dB(s)
ds · B(s). It suffices to show
a(s) = c(s) = 0 for any s ∈ I.
d 2
Since B(s) is unit, one can easily see that c(s) ≡ 0 by considering ds |B| (c.f. Lemma
6.7). To show a(s) ≡ 0, we consider the fact that:
dB
Remark 6.19. By Lemma 6.17, the vector ds and N are parallel. Combining with the
fact that N is unit, one can see easily that:
dB dB
|τ (s)| = |N| cos 0 = .
ds ds
Therefore, the torsion can be regarded as a signed dB
ds which measures the rate that the
osculating plane rotates as s increases (see Figure 6.3). The negative sign appeared in
the definition is a historical convention.
Figure 6.3. Torsion measures how fast the osculating plane changes along a curve
Example 6.20. Consider the curve γ(s) = cos √s2 , sin √s2 , √s2 which is the helix
appeared in Example 6.14. The normal and binormal were already computed:
γ 00 (s)
s s
N(s) = 00 = − cos √ , − sin √ , 0
|γ (s)| 2 2
1 s 1 s 1
B(s) = √ sin √ , − √ cos √ , √ .
2 2 2 2 2
Taking the derivative, we get:
dB 1 s 1 s
= cos √ , sin √ , 0 .
ds 2 2 2 2
Therefore, the torsion of the curve is:
dB 1
τ (s) = − ·N= .
ds 2
Exercise 6.9. Consider the curve γ(t) = (a cos t, a sin t, bt) where a and b are
positive constants. Find its torsion τ (s) as a function of the arc-length parameter s.
The torsion of a non-degenerate curve γ(t) can be difficult to compute from the
definition since it involves finding an explicit arc-length parametrization. Fortunately,
just like the curvature, there is a formula for computing torsion.
Proposition 6.21. Let γ(t) : I → R3 be a non-degenerate curve, then the torsion of the
curve is given by:
(γ 0 (t) × γ 00 (t)) · γ 000 (t)
τ (t) = 2 .
|γ 0 (t) × γ 00 (t)|
Exercise 6.12. The purpose of this exercise is to give a proof of Proposition 6.21. As
γ(t) is a (regular) non-degenerate curve, there exist an arc-length parametrization
γ(s) := γ(t(s)) and a Frenet-Serret Frame {T(s), N(s), B(s)} at every point on the
curve. With a little abuse of notations, we denote κ(s) := κ(t(s)) and τ (s) := τ (t(s)).
(a) Show that
(γ 0 (s) × γ 00 (s)) · γ 000 (s)
τ (s) = .
κ(s)2
(b) Using (6.3) in the proof of Proposition 6.8, show that
3
ds
γ 000 (t) = γ 000 (s) + v(s)
dt
where v(s) is a linear combination of γ 0 (s) and γ 00 (s) for any s.
(c) Hence, show that
(γ 0 (t) × γ 00 (t)) · γ 000 (t)
(γ 0 (s) × γ 00 (s)) · γ 000 (s) = 6 .
|γ 0 (t)|
[Hint: use (6.4) in the proof of Proposition 6.8.]
(d) Finally, complete the proof of Proposition 6.21. You may use the curvature
formula proved in Proposition 6.8.
Exercise 6.13. Compute the torsion τ (t) for the ellipsoidal helix:
γ(t) = (a cos t, b sin t, ct)
where a and b are positive and c is non-zero.
176 6. Geometry of Curves
The Fundamental Theorem of Space Curves answers both questions above. Using the
classic existence and uniqueness theorems in Ordinary Differential Equations (ODEs), one
can give an affirmative answer to the above existence question – yes, such a curve exists –
and an “almost” affirmative answer to the uniqueness question – that is, although the
curves γ(s) and γ̄(s) may not be identical, one can be transformed from another by a rigid
body motion in R3 . The proof of this theorem is a good illustration of how Differential
Equations interact with Differential Geometry – nowadays a field called Geometric Analysis.
6.2.1. Existence and Uniqueness of ODEs. A system of ODEs (or ODE system) is
a set of one or more ODEs. The general form of an ODE system is:
where t is the independent variable, xi (t)’s are unknown functions, and fj ’s are prescribed
functions of (x1 , . . . , xn , t) from Rn × I → R.
An ODE system with a given initial condition, such as (x1 (0), . . . , xn (0)) =
(a1 , . . . , an ) where ai ’s are constants, is called an initial-value problem (IVP).
We first state a fundamental existence and uniqueness theorem for ODE systems:
6.2. Fundamental Theorem of Space Curves 177
6.2.3. Fundamental Theorem. We now state the main theorem of this section:
Theorem 6.23 (Fundamental Theorem of Space Curves). Given any smooth positive
function α(s) : I → (0, ∞), and a smooth real-valued function β(s) : I → R, there exists
an arc-length parametrized, non-degenerate curve γ(s) : I → R3 such that its curvature
κ(s) ≡ α(s) and its torsion τ (s) ≡ β(s).
Moreover, if γ̄(s) : I → R3 is another arc-length parametrized, non-degenerate curve
whose curvature κ̄(s) ≡ α(s) and torsion τ̄ (s) ≡ β(s), then there exists a 3 × 3 constant
matrix A with AT A = I, and a constant vector p, such that γ̄(s) = Aγ(s) + p for any
s ∈ I.
Step 2: Show that there exists a curve γ(s) whose Frenet-Serret Frame is given by
T(s) = ê1 (s), N(s) = ê2 (s) and B(s) = ê3 (s). Consequently, from the system (6.7),
one can claim γ(s) is a curve that satisfies the required conditions.
Step 3: Prove the uniqueness part of the theorem.
Step 1: To begin, let’s consider the ODE system with unknowns ê1 , ê2 and ê3 :
0
ê1 (s) 0 α(s) 0 ê1 (s)
(6.7) ê2 (s) = −α(s) 0 β(s) ê2 (s)
ê3 (s) 0 −β(s) 0 ê3 (s)
which is an analogous system to the Frenet-Serret System (6.6). Impose the initial
conditions:
ê1 (0) = î, ê2 (0) = ĵ, ê3 (0) = k̂.
Recall that α(s) and β(s) are given to be smooth (in particular, continuously differ-
entiable). By Theorem 6.22, there exists a solution {ê1 (s), ê2 (s), ê3 (s)} defined on a
maximal interval s ∈ (T− , T+ ) that satisfies the system with the above initial conditions.
Note that {ê1 (s), ê2 (s), ê3 (s)} is orthonormal initially at s = 0, we claim it remains
so as long as solution exists. To prove this, we first derive (see Exercise 6.14):
ê1 · ê1 0 0 0 2α 0 0 ê1 · ê1
ê2 · ê2 0 0 −2α 2β
0 0 ê2 · ê2
d ê3 · ê3 0
0 0 0 −2β 0 ê3 · ê3
(6.8) =
ds ê1 · ê2 −α α 0
0 0 β ê1 · ê2
ê2 · ê3 0 −β β 0 0 −α ê2 · ê3
ê3 · ê1 0 0 0 −β α 0 ê3 · ê1
Regarding êi · êj ’s are unknowns, (6.8) is a linear ODE system of 6 equations with
initial conditions:
(ê1 · ê1 , ê2 · ê2 , ê3 · ê3 , ê1 · ê2 , ê2 · ê3 , ê3 · ê1 )s=0 = (1, 1, 1, 0, 0, 0)
It can be verified easily that the constant solution (1, 1, 1, 0, 0, 0) is indeed a solution
to (6.8). Therefore, by the uniqueness part of Theorem 6.22, we must have
(ê1 · ê1 , ê2 · ê2 , ê3 · ê3 , ê1 · ê2 , ê2 · ê3 , ê3 · ê1 ) = (1, 1, 1, 0, 0, 0)
for any s ∈ (T− , T+ ). In other words, the frame {ê1 (s), ê2 (s), ê3 (s)} is orthonormal as
long as solution exists.
Consequently, each of {ê1 (s), ê2 (s), ê3 (s)} remains bounded, and by last statement
of Theorem 6.22, this orthonormal frame can be extended so that it is defined for all
s ∈ I.
Step 2: Using the frame ê1 (s) : I → R3 obtained in Step 1, we define:
Z s
γ(s) = ê1 (s) ds.
0
Evidently, γ(s) is a curve starting from the origin at s = 0. Since ê1 (s) is continuous, γ(s)
is well-defined on I and by the Fundamental Theorem of Calculus, we get:
T(s) := γ 0 (s) = ê1 (s)
which is a unit vector for any s ∈ I. Therefore, γ(s) is arc-length parametrized.
Next we verify that γ(s) is the curve require by computing its curvature and torsion.
By (6.8),
γ 00 (s) = ê01 (s) = α(s)ê2 (s)
6.2. Fundamental Theorem of Space Curves 179
The existence part of Theorem 6.23 only shows a curve with prescribed curvature
and torsion exists, but it is in general difficult to find such a curve explicitly. While the
existence part does not have much practical use, the uniqueness part has some nice
corollaries.
First recall that a helix is a curve of the form γa,b (t) = (a cos t, a sin t, bt) where
a 6= 0 and b can be any real number. It’s arc-length parametrization is given by:
s s bs
γa,b (s) = a cos √ , a sin √ , √ .
a2 + b2 a2 + b2 a2 + b2
It can be computed that its curvature and torsion are both constants:
a
κa,b (s) ≡ 2
a + b2
b
τa,b (s) ≡ 2 .
a + b2
Conversely, given two constants κ0 > 0 and τ0 ∈ R, by taking a = κ2κ+τ 0
2 and
0 0
τ0
b = κ2 +τ 2 , the helix γa,b (s) with this pair of a and b has curvature κ0 and torsion τ0 .
0 0
Hence, the uniqueness part of Theorem 6.23 asserts that:
Corollary 6.24. A non-degenerate curve γ(s) has constant curvature and torsion if and
only if γ(s) is congruent to one of the helices γa,b (s).
Remark 6.25. Two space curves γ(s) and γ e(s) are said to be congruent if there exists a
3 × 3 orthogonal matrix A and a constant vector p ∈ R3 such that γ e(s) = Aγ(s) + p. In
simpler terms, one can obtain γ
e(s) by rotating and translating γ(s).
6.3. Plane Curves 181
One can easily check that {T(s), JT(s)} is an orthonormal frame of R2 for any s ∈ I.
Let’s call this the TN-Frame of the curve. We will work with the TN-Frame in place of the
Frenet-Serret Frame for plane curves. The reasons for doing so are two-folded. For one
thing, the normal frame. JT(s) is well-defined for any s ∈ I even though κ(s) is zero for
some s ∈ I. Hence, one can relax the non-degeneracy assumption here. For another, we
can introduce the signed curvature k(s):
Note that T(s) is unit, so by Lemma 6.7 we know T(s) and T0 (s) are always or-
thogonal and hence it is either in or against the direction of JT(s). Therefore, we
have
|k(s)| = |T0 (s)| |JT(s)| = |γ 00 (s)| = κ(s).
The sign of k(s) is determined by whether T0 and ĵT are along or against each other (see
Figure 6.4).
Example 6.27. Let’s compute the signed curvature of the unit circle
γ(s) = (cos εs, sin εs)
182 6. Geometry of Curves
Exercise 6.15. Consider a plane curve γ(s) parametrized by arc-length. Let θ(s) be
the angle between the x-axis and the unit tangent vector T(s). Show that:
T0 (s) = θ0 (s)JT(s) and k(s) = θ0 (s).
Exercise 6.16. [dC76, P.25] Consider a plane curve γ(s) parametrized by arc-length.
Suppose |γ(s)| is maximum at s = s0 . Show that:
1
|k(s0 )| ≥ .
|γ(s0 )|
Theorem 6.28 (Fundamental Theorem of Plane Curves). Given any smooth real-valued
function α(s) : I → R, there exists a regular plane curve γ(s) : I → R2 such that its signed
curvature k(s) ≡ α(s). Moreover, if γ̄(s) : I → R2 is another regular plane curve such
that its signed curvature k̄(s) ≡ α(s), then there exists a 2 × 2 orthogonal matrix A and a
constant vector p ∈ R2 such that γ̄(s) ≡ Aγ(s) + p.
Exercise 6.17. Prove Theorem 6.28. Although the proof is similar to that of Theorem
6.23 for non-degenerate space curves, please do not use the latter to prove the former
in this exercise. Here is a hint on how to begin the proof: Consider the initial-value
problem
ê0 (s) = α(s) J ê(s)
ê(0) = î
6.3. Plane Curves 183
Exercise 6.18. Using Theorem 6.28, show that a regular plane curve has constant
signed curvature if and only if it is a straight line or a circle
Exercise 6.19. [Küh05, P.50] Find an explicit plane curve γ(s) such that the signed
curvature is given by k(s) = √1s .
The following is a celebrated result that relates the local property (i.e. signed
curvature) to the global property (topology) of simple closed curves:
Theorem 6.30 (Hopf). For any arc-length parametrized, simple closed curve γ(s) :
[0, L] → R2 such that γ 0 (0) = γ 0 (L), we must have:
Z L
k(s) ds = ±2π.
0
The original proof was due to Hopf. We will not discuss Hopf’s original proof in this
course, but we will prove a weaker result, under the same assumption as Theorem 6.30,
that Z L
k(s) ds = 2πn
0
for some integer n.
Let {T(s), JT(s)} be the TN-frame of γ(s). Since T(s) is unit for any s ∈ [0, L], one
can find a smooth function θ(s) : [0, L] → R such that
T(s) = (cos θ(s), sin θ(s))
for any s ∈ [0, L]. Here θ(s) can be regarded as 2kπ + angle between T(s) and î. In
order to ensure continuity, we allow θ(s) to take values beyond [0, 2π].
Then JT(s) = (− sin θ(s), cos θ(s)), and so by the definition of k(s), we get:
k(s) = T0 (s) · JT(s)
= (−θ0 (s) sin θ(s), θ0 (s) cos θ(s)) · (− sin θ(s), cos θ(s))
= θ0 (s).
Therefore, the total curvature is given by:
Z L Z L
k(s) ds = θ0 (s) ds = θ(L) − θ(0).
0 0
Since it is assumed that T(0) = T(L) in Theorem 6.30, we have
θ(L) ≡ θ(0) (mod 2π)
and so we have: Z L
k(s) ds = 2πn
0
for some integer n.
Chapter 7
Geometry of Euclidean
Hypersurfaces
Henri Poincaré
185
186 7. Geometry of Euclidean Hypersurfaces
∂Fα
Each vector is a tangent to Σn at the based point p = Fα (u1α , · · · , unα ).
∂uiα (u1 ,··· ,un )
α α
With a bit abuse of notations and for simplicity, we will from now on denote
∂Fα ∂Fα
i
(p) :=
∂uα ∂uiα (u1α ,··· ,un
α)
where p = Fα (u1α , · · · , unα ), to emphasize that p is the based point of the tangent vector.
By condition (3) in Definition 7.1, we know that the following vector space
∂Fα ∂Fα
Tp Σn := span (p), · · · , (p)
∂u1α ∂unα
has dimension n. This is called the tangent space at p ∈ Σn .
Example 7.2. The graph Σf of any smooth function f : Rn → R is a regular hypersurface
in Rn+1 . One can parametrize Σf by a single parametrization:
F : Rn → Σf
(x1 , · · · , xn ) 7→ x1 , · · · , xn , f (x1 , · · · , xn )
By straight-forward computations, we have:
∂F ∂f
= êi + ên+1
∂xi ∂xi
n on
∂F
where {ê1 , · · · , ên+1 } is the standard basis vectors in Rn+1 . It is clear that ∂xi are
i=1
linearly independent.
Example 7.3. The n-dimensional unit sphere:
( n+1
)
X
n n+1
S := (x1 , · · · , xn+1 ) ∈ R : x2i =1
i=1
Exercise 7.1. Fill in the omitted detail in Examples 7.2 and 7.3.
Exercise 7.2. Show that any regular hypersurface in Rn+1 is a smooth manifold,
and also a smooth submanifold of Rn+1 .
Regular hypersurfaces are the higher dimensional generalization of regular surfaces
discussed in Chapter 1. As such many important results about regular surfaces can carry
over naturally to regular hypersurfaces. We state the results below and leave the proofs
as exercises.
Theorem 7.4 (c.f. Theorem 1.6). Let g(x1 , · · · , xn+1 ) : Rn+1 → R be a smooth function.
Consider a non-empty level set g −1 (c) where c is a constant. If ∇g(p) 6= 0 at all points
p ∈ g −1 (c), then the level set g −1 (c) is a regular hypersurface in Rn+1 .
7.1. Regular Hypersurfaces in Euclidean Spaces 187
Proposition 7.5 (c.f. Proposition 1.8). Assume all given conditions stated in Theorem
7.4. Furthermore, suppose F is a bijective map from an open set U ⊂ Rn to an open set
O ⊂ Σ := g −1 (c) which satisfies conditions (1) and (3) in Definition 7.1. Then, F satisfies
condition (2) as well and hence is a smooth local parametrization of g −1 (c).
Proposition 7.6 (c.f. Proposition 1.11). Let Σ ⊂ Rn+1 be a regular surface, and
Fα : Uα → M and Fβ : Uβ → M be two smooth local parametrizations of Σ with
overlapping images, i.e. W := Fα (Uα ) ∩ Fβ (Uβ ) 6= ∅. Then, the transition maps defined
below are also smooth maps:
(Fβ−1 ◦ Fα ) : Fα−1 (W) → Fβ−1 (W)
(Fα−1 ◦ Fβ ) : Fβ−1 (W) → Fα−1 (W)
Exercise 7.3. Prove Theorem 7.4, Proposition 7.5, and Proposition 7.6.
188 7. Geometry of Euclidean Hypersurfaces
Definition 7.7 (First Fundamental Form). The first fundamental form of a regular
hypersurface Σn ⊂ Rn+1 is a 2-tensor g on Σ with local expression g = gij dui ⊗ duj
where gij is given by:
∂F ∂F
gij = , ,
∂ui ∂uj
where h·, ·i denotes the usual dot product on Rn+1 .
Exercise 7.4. Show that gij dui ⊗ duj is independent of local coordinates, i.e. if
G(vα ) is another local parametrization and denote
∂G ∂G
geαβ = , ,
∂vα ∂vβ
then
geαβ dv α ⊗ dv β = gij dui ⊗ duj .
Example 7.8. Let S2 be the unit 2-sphere and F be the following local parametrization:
F (u, v) = (sin u cos v, sin u sin v, cos u), (u, v) ∈ (0, π) × (0, 2π)
By direct computations, we have:
∂F
= (cos u cos v, cos u sin v, − sin u)
∂u
∂F
= (− sin u sin v, sin u cos v, 0)
∂v
∂F ∂F ∂F ∂F ∂F ∂F ∂F ∂F
g , = · =1 g , = · =0
∂u ∂u ∂u ∂u ∂u ∂v ∂u ∂v
∂F ∂F ∂F ∂F ∂F ∂F ∂F ∂F
g , = · =0 g , = · = sin2 u
∂v ∂u ∂v ∂u ∂v ∂v ∂v ∂v
Therefore, we get its first fundamental equals
g = du ⊗ du + sin2 u dv ⊗ dv.
7.2. Fundamental Forms and Curvatures 189
Exercise 7.6. Show that the first fundamental form of the graph Σf in Example 7.2
is given by:
∂f ∂f
g = δij + dxi ⊗ dxj = δij dxi ⊗ dxj + df ⊗ df.
∂xi ∂xj
As gij = gji , the tensor g is symmetric. As such the tensor notation gij dui ⊗ duj is
often written simply as gij dui duj . For instance, the first fundamental form of the unit
sphere in Example 7.8 can be expressed as:
g = du2 + sin2 u dv 2
where du2 is interpreted as (du)2 = du du, not d(u2 ).
Another way to represent the first fundamental form is by the matrix:
g11 g12 · · · g1n
g21 g22 · · · g2n
[g] := . .. .. .
.. . .
gn1 gn2 · · · gnn
∂F ∂F ∂F ∂F
It is a symmetric matrix since g ∂u i
, ∂uj = g ∂uj ∂ui .
,
∂F ∂F
Given two tangent vectors Y = Y i ∂u i
and Z = Z i ∂u i
in Tp Σ, the value of g(Y, Z) is
related to the entries of [g] in the following way:
∂F ∂F
g(Y, Z) = g Y i , Zj = Y i Z j gij
∂ui ∂uj
g11 g12 · · · g1n 1
Z
1 g21 g22 · · · g2n
n ..
= Y ··· Y . .. .. .
.. . .
Zn
gn1 gn2 · · · gnn
Note that the matrix [g] depends on local coordinates although the tensor g does not.
As computed in Example 7.8, the matrix [g] of the unit sphere (with respect the
parametrization F used in the example) is given by:
1 0
[g] := 2
0 sin u
Evidently, it is a diagonal matrix. Try to think about the geometric meaning of [g] being
diagonal.
We will see in subsequent sections that g “encodes” crucial geometric information
such as curvatures. There are also some familiar geometric quantities, such as length
and area, which are related to the first fundamental form g.
Consider a curve γ on a regular hypersurface Σn ⊂ Rn+1 parametrized by F (ui ).
Suppose the curve can be parametrized by γ(t), a < t < b, then from calculus we know
the arc-length of the curve is given by:
Z b
L(γ) = |γ 0 (t)| dt
a
In fact one can express this above in terms of g. The argument is as follows:
Suppose γ(t) has local coordinates coordinates (γ i (t)) such that F (γ i (t)) = γ(t).
Using the chain rule, we then have:
∂F dγ i
γ 0 (t) = .
∂ui dt
190 7. Geometry of Euclidean Hypersurfaces
p
Recall that |γ 0 (t)| = hγ 0 (t), γ 0 (t)i and that γ 0 (t) lies on Tp M , we then have:
p
|γ 0 (t)| = g (γ 0 (t), γ 0 (t))
We can then express it in terms of the matrix components gij ’s:
dγ i dγ j
(7.1) g (γ 0 (t), γ 0 (t)) = gij
dt dt
where gij ’s are evaluated at the point γ(t). Therefore, the arc-length can be expressed in
terms of the first fundamental form by:
Z bp Z br
dγ i dγ j
L(γ) = g (γ 0 (t), γ 0 (t)) dt = gij dt
a a dt dt
Another familiar geometric quantity which is also related to g is the area of a surface.
For simplicity, we focus on dimension 2 first. Suppose a regular surface Σ can be almost
everywhere parametrized by F (u, v) with (u, v) ∈ D ⊂ R2 where D is a bounded domain,
the area of this surface is given by:
ZZ
∂F ∂F
A(M ) = × dudv
D ∂u ∂v
Therefore, the length the curve γ(t) := F (γ 1 (t), · · · , γ n (t)) can be computed by
integrating the square root of:
n 2 n
dγ i dγ j X dγ i X ∂f dγ i ∂f dγ j
gij = +
dt dt i=1
dt i,j=1
∂ui dt ∂uj dt
n 2 2
dγ i
X d
= + f (γ(t)) .
i=1
dt dt
1Note that ∇f is an eigenvector with eigenvalue 1 + |∇f |2 , and (∇f )(∇f )T has rank 1.
192 7. Geometry of Euclidean Hypersurfaces
Given a regular hypersurface, there are always two choices of normal vector at each
point. For a sphere, once the normal vector direction is chosen, it is always consistent
with your choice when we move the normal vector across the sphere. That is, when you
draw a closed path on the sphere and see how the unit normal vector varies along the
path, you will find that the unit normal remains the same when you come back to the
original point. We call it an orientable hypersurface if there exists a continuous choice
of unit normal vector across the whole hypersurface. See Chapter 4 for more thorough
discussions on orientability.
When Σ is an orientable regular hypersurface, the chosen unit normal vector ν can
then be regarded as a map. The domain of ν is Σ. Since ν is unit, the codomain can be
taken to be the unit sphere Sn . We call this map as:
As computed in Example 7.10, the Gauss map ν for the unit sphere S2 is given by F
(assuming the outward-pointing convention is observed). It is not difficult to see that
the Gauss map ν for a sphere with radius R centered the origin in R3 is given by R1 F .
Readers should verify this as an exercise.
For a plane Π, the unit normal vector at each point is the same. Therefore, the Gauss
map ν(p) is a constant vector independent of p.
A unit cylinder with z-axis as its central axis can be parametrized by:
F (u, v) = (cos u, sin u, v), (u, v) ∈ (0, 2π) × R.
By straight-forward computations, one can get:
∂F ∂F
× = (cos u, sin u, 0)
∂u ∂v
which is already unit. Therefore, the Gauss map of the cylinder is given by:
ν(u, v) = (cos u, sin u, 0).
2
The image of ν in S is the equator.
It is not difficult to see that the image of the Gauss map ν, which is a subset of S2 , is
related to how “spherical” or “planar” the surface looks. The smaller the image, the more
planar it is.
The curvature of a regular curve is a scalar function κ(p). Since a curve is one
dimensional, we can simply use one single value to measure the curvature at each point.
However, a regular hypersurface has arbitrary dimensions and hence has higher degree
of freedom than curves. It may bend by a different extent along different direction. As
such, there are various notions of curvatures for regular hypersurfaces. We first talk
about the normal curvature, which is fundamental to many other notions of curvatures.
Let Σ be an orientable regular hypersurface with its Gauss map denoted by ν. At each
point p ∈ Σ, we pick a unit tangent vector ê in Tp Σ. Heuristically, the normal curvature
at p measures the curvature of the surface along a direction ê. Precisely, we define:
Remark 7.13. Since kn (p, ê) is defined using the Gauss map ν, which always comes with
two possible choice for any orientable regular surface, the normal curvature depends on
the choice of the Gauss map ν. If the opposite unit normal is chosen to be the Gauss map,
the normal curvature will differ by a sign.
We will first make sense of normal curvatures through elementary examples, then
we will prove a general formula for computing normal curvatures.
Example 7.14. Let P be any plane in R3 . For any point p ∈ P and unit tangent ê ∈ Tp P ,
the plane Πp (ν, ê) must cut through P along a straight-line γ. Since γ has curvature 0,
we have:
kn (p, ê) = 0
for any p ∈ P and ê ∈ Tp P . See Figure 7.2a.
Example 7.15. Let S2 (R) be the sphere with radius R centered at the origin in R3 with
Gauss map ν taken to be inward-pointing. For any point p ∈ S2 and ê ∈ Tp S2 (R), the
plane Πp (ν, ê) cuts S2 (R) along a great circle (with radius R). Since a circle with radius
R has constant curvature R1 , we have:
1
kn (p, ê) =
R
for any p ∈ S2 (R) and ê ∈ Tp S2 (R). See Figure 7.2b.
194 7. Geometry of Euclidean Hypersurfaces
(c) cylinder
Example 7.16. Let M be the (infinite) cylinder of radius R with x-axis as the central axis
with outward-pointing Gauss map ν. Given any p ∈ M , if êx is the unit tangent vector
at p parallel to the x-axis, then the Πp (ν, êx ) cuts the cylinder M along a straight-line.
Therefore, we have:
kn (p, êx ) = 0
for any p ∈ M . See the blue curve in Figure 7.2c.
On the other hand, if êyz is a horizontal unit tangent vector at p, then Πp (ν, êyz ) cuts
M along a circle with radius R. Therefore, we have:
1
kn (p, êyz ) = −
R
for any p ∈ M . See the red curve in Figure 7.2c. Note that the tangent vector of the
curve is moving away from the outward-pointing ν. It explains the negative sign above.
For any other choice of unit tangent ê at p, the plane Πp (ν, ê) cuts the cylinder along
an ellipse, so the normal curvature along ê may vary between 0 and − R1 .
In the above examples, the normal curvatures kn (p, ê) are easy to find since the
curve of intersection between Πp (ν, ê) and the surface is either a straight line or a circle.
Generally speaking, the curve of intersection may be of arbitrary shape such as an ellipse,
and sometimes it is not even easy to identify what curve it is. Fortunately, it is possible
to compute kn (p, ê) for any given unit tangent ê in a systematic way. Next we derive an
expression of kn (p, ê), which will motivate the definition of the second fundamental form
and Weingarten’s map (also known as the shape operator).
7.2. Fundamental Forms and Curvatures 195
Proposition 7.17. Let Σ be the regular hypersurface in Rn+1 with Gauss map ν. Fix
p ∈ Σ and a unit vector ê ∈ Tp Σ, then the normal curvature of Σ at p along ê is given by:
kn (p, ê) = − hê, Dê νi .
Furthermore, suppose ê can be locally expressed as
∂F
ê = X i ,
∂ui
then we have
2
∂F ∂ν i j ∂ F
kn (p, ê) = − , X X = , ν X iX j .
∂ui ∂uj ∂ui ∂uj
Proof. Let γ be the intersection curve between the plane Πp (ν, ê) and Σ. We parametrize
γ by arc-length s ∈ (−ε, ε) such that γ(0) = p and |γ 0 (s)| ≡ 1 on (−ε, ε). Then, γ 0 (s) is
orthogonal to ν(γ(s)), and so
The last step follows from the fact that γ 0 (0) = ê.
The local expression follows immediately from the fact that
∂ν
DX i ∂F ν = X i D ∂F ν = X i .
∂ui ∂ui ∂ui
Given that kn (p, ê) depends on the unit direction ê ∈ Tp Σ, it is then natural to ask
when kn (p, ê) achieves the maximum and minimum among all unit vectors ê in Tp Σ. It is
a simple optimization problem of critical points of the h(ê, ê) (as a function of ê) subject
to the condition g(ê, ê) = 1. We call the critical values of h(ê, ê) subject to g(ê, ê) = 1 to
be principal curvatures and we have the following important result:
Proposition 7.22. Given a regular hypersurface Σ in Rn+1 and fix a point p ∈ Σ, the
principal curvatures are eigenvalues of the linear map:
Sp : Tp Σ → Tp Σ
∂F ∂F
7→ g jk hki .
∂ui ∂uj
The map Sp is called the Weingarten’s map, or the shape operator.
∂F 2
Proof. Denote ê = X i ∂u i
, then kn (p, ê) = h(ê, ê) = hij X i X j and |ê| = gij X i X j . To
determine the principal curvatures, we use Lagrange’s multiplier to find the critical
values of hij X i X j under the constraint gij X i X j = 1. Here we treat (X 1 , · · · , X n ) as
the variables, while gij and hij are regarded as constants since they do not depend on ê.
7.2. Fundamental Forms and Curvatures 197
Therefore, if kn (p, ê) achieves its maximum and minimum among all ê ∈ Tp Σ, then
kn (p, ê) is an eigenvalue of [g]−1 [h], which is the matrix representation of S with respect
to local coordinates (u1 , · · · , un ) as desired
It turns out that when dim Σ = 2, the Gauss curvature K depends only on the first
fundamental form even though it is defined using the second fundamental form as well.
It is a famous theorem by Gauss commonly known as Theorema Egregium.
198 7. Geometry of Euclidean Hypersurfaces
Proof. At a critical p of f , we have ∇f (p) = 0. From Examples 7.9 and 7.21, we have
computed:
∂f ∂f
gij (p) = δij + (p) (p) = δij
∂ui ∂uj
| {z }
=0
∂2f
∂ui ∂uj (p) ∂2f
hij (p) = q = (p)
2 ∂ui ∂uj
1 + |∇f (p)|
Note that the Gauss map ν was taken to be upward-pointing in Example 7.21, as required
in this proposition.
Therefore, we have:
2
det[h] ∂ f 2
K(p) = (p) = det (p) = f11 f22 − f12 (p)
det[g] ∂ui ∂uj
1 ij 1 ∂2f 1
H(p) = g (p)hij (p) = δ ij (p) = (f11 + f22 ) (p)
2 2 ∂ui ∂uj 2
From the second derivative test in multivariable calculus, given a critical point p, if
2
f11 f22 − f12 > 0 and f11 + f22 > 0 at p, then p is a local minimum of f . The other cases
can be proved similarly using the second derivative test.
Given any regular surface Σ (not necessarily the graph of a function) and any point
p ∈ Σ, one can apply a rigid-motion motion Φ : R3 → R3 so that Tp Σ is transformed into
a horizontal plane. Then, the new surface Φ(Σ) becomes locally a graph of a function f
near the point Φ(p). Recall that the Gauss curvatures of p and Φ(p) are the same as given
by Exercise 7.8. If K(p) > 0 (and hence K(Φ(p)) > 0), then Proposition 7.25 asserts that
7.2. Fundamental Forms and Curvatures 199
Φ(p) is a local maximum or minimum of the function f and so the surface Φ(Σ) is locally
above or below the tangent plane at Φ(p). In other words, near p the surface Σ is locally
on one side of the the tangent plane Tp Σ. On the other hand, if K(p) < 0 then no matter
how close to p the surface Σ would intersect Tp Σ at points other than p.
Definition 7.26 (Surfaces of Revolution). Consider the curve γ(t) = (x(t), 0, z(t)),
where t ∈ (a, b), on the xz-plane such that x(t) ≥ 0 for any t ∈ (a, b). The surface of
revolution generated by γ is obtained by revolving γ about the z-axis, and it can be
parametrized by:
F (t, θ) = (x(t) cos θ, x(t) sin θ, z(t)), (t, θ) ∈ (a, b) × [0, 2π].
Under the condition on γ(t) = (x(t), 0, z(t)) that its surface of revolution is smooth,
one can easily compute that the first fundamental form is given by:
0 2 2
(x ) + (z 0 ) 0
(7.8) [g] = (matrix notation)
0 x2
h i
2 2
g = (x0 ) + (z 0 ) dt2 + x2 dθ2 (tensor notation)
∂F
× ∂F
and the second fundamental form with respect to the Gauss map ν := ∂t ∂θ
is given
| ∂F
∂t ∂θ |
× ∂F
by:
0 00
x z − x00 z 0
1 0
(7.9) [h] = q (matrix notation)
2 2 0 x z0
(x0 ) + (z 0 )
1 0 00
(x z − x00 z 0 ) dt2 + x z 0 dθ2
h= q (tensor notation)
2 2
(x0 ) + (z 0 )
200 7. Geometry of Euclidean Hypersurfaces
Exercise 7.10. Verify that the first and second fundamental forms of a surface of
revolution with parametrization
F (t, θ) = (x(t) cos θ, x(t) sin θ, z(t)), (t, θ) ∈ (a, b) × [0, 2π]
are given as in (7.8) and (7.9).
As both [g] and [h] are diagonal matrices, it is evident that the principal curvatures,
i.e. the eigenvalues of [g]−1 [h], are:
x0 z 00 − x00 z 0 x0 z 00 − x00 z 0
k1 = h i3/2 = 3
2
(x0 ) + (z 0 )
2 |γ 0 |
z0 z0
k2 = q =
2 2 x |γ 0 |
x (x0 ) + (z 0 )
Note that here we are not using the convention that k1 ≤ k2 as in before, since there is
no clear way to tell which eigenvalue is larger.
Therefore, the Gauss and mean curvatures are given by:
(x0 z 00 − x00 z 0 ) z 0
(7.10) K = k1 k2 = 4
x |γ 0 |
!
1 1 x0 z 00 − x00 z 0 z0
(7.11) H = (k1 + k2 ) = 3 +
2 2 |γ 0 | x |γ 0 |
Example 7.27. Let’s verify that the round sphere (of any radius) has indeed constant
Gauss and mean curvatures. Parametrize the sphere by:
F (t, θ) = (R sin t cos θ, R sin t sin θ, R cos t),
i.e. taking x(t) = R sin t and z(t) = R cos t. By (7.10), then the Gauss curvature is clearly
given by:
(R cos t)(−R cos t) − (−R sin t)(−R sin t) (−R sin t) 1
K= = 2.
(R sin t) · R4 R
By (7.11), the mean curvature is given by:
1 −R2
−R sin t 1
H= + =− .
2 R3 (R sin t) · R R
Exercise 7.11. Compute the Gauss and mean curvatures of a round torus using
(7.10) and (7.11).
7.3. Theorema Egregium 201
∂ ∂F ∂2F
Dγ 0 X = = .
∂ui ∂uj ∂ui ∂uj
∂F
In general, suppose X = X i ∂u i
and given any curve γ(t) on Σ locally expressed as
(u1 (t), · · · , un (t)), then by the chain rule we have:
d ∂X dui
Dγ 0 X = X(γ(t)) =
dt ∂u dt
i
∂ ∂F dui
= Xj
∂ui ∂uj dt
∂X j ∂F ∂2F
dui
= + Xj
∂ui ∂uj ∂ui ∂uj dt
∂X j ∂F
Note that under a fixed local coordinate system (u1 , · · · , un ), the quantities ∂ui ∂uj +
j ∂2F dui
X are uniquely determined by the vector field X whereas
∂ui ∂uj are uniquely dt
determined by the tangent vector γ 0 of the curve. Now given another vector field
∂F
Y = Y i ∂ui
, then one can define
DY X(p) := Dγ 0 X(p)
202 7. Geometry of Euclidean Hypersurfaces
where γ is any curve on Σ which solves the ODE γ 0 (t) = Y (γ(t)) and γ(0) = p. The
Existence Theorem of ODEs guarantees there is such a curve γ that flows along Y . Locally,
DY X can be expressed as:
∂X j ∂F ∂2F
(7.12) DY X = + Xj Y i.
∂ui ∂uj ∂ui ∂uj
Definition 7.28 (Covariant Derivatives). Let Σ be a regular surface with Gauss map ν,
and γ(t) be a smooth curve on Σ. Given two vector fields X and Y on Σ, we define the
covariant derivative of X at p ∈ M along Y to be
T
∇Y X(p) := (DY X(p)) = DY X − hDY X, νi ν (p).
T
Here (DY X(p)) represents the projection of DY X(p) onto the tangent space Tp Σ.
Using the fact that hDY X, νi = −hX, DY νi = h(X, Y ) = hij X i Y j and (7.12), we
can derive the local expression for ∇Y X:
X ∂X j ∂F 2
∂ F
X
(7.13) ∇Y X = + Xj Y i − hij X i Y j ν
i,j
∂ui ∂uj ∂ui ∂uj i,j
| {z } | {z }
DY X hDY X,νiν
Suppose X, X, e Y and Ye are vector fields and ϕ is a smooth scalar functions. One
can verify that the following properties hold:
(a) ∇ϕY X = ϕ∇Y X
(b) ∇Y (ϕX) = (∇Y ϕ) X + ϕ∇Y X
(c) ∇Y +Ye X = ∇Y X + ∇Ye X
(d) ∇Y (X + X)
e = ∇Y X + ∇Y X
e
∂F ∂F
According to (7.12) and (7.13), given any vector fields X = X i ∂ui
and Y = Y i ∂u i
,
∂2F
the second derivatives determine both DY X and ∇Y X. We are going to express
∂ui ∂uj n o
∂2F ∂F ∂F
in terms of this tangent-normal basis ∂u
∂ui ∂uj 1
, · · · , ∂un
, ν of Rn+1 . From (7.13),
we have:
∂2F ∂F ∂F
= D ∂F = ∇ ∂F + hij ν.
∂ui ∂uj ∂u j ∂ui ∂u j ∂ui
∂F
We denote the coefficients of the tangent vector ∇ ∂F ∂ui by the following Christoffel
∂uj
symbols:
∂F ∂F
(7.14) ∇ ∂F = Γkij .
∂uj ∂ui ∂uk
The Christoffel symbols can be shown to be depending only on the first fundamental
form g. We will use it to prove that Gauss curvature depends also only on g but not on h.
7.3. Theorema Egregium 203
Proposition 7.29. In a local coordinate system (u1 , · · · , un ), the Christoffel symbols Γkij ’s
can be locally expressed in terms of the first fundamental form as:
k 1 kl ∂gjl ∂gil ∂gij
(7.15) Γij = g + − .
2 ∂ui ∂uj ∂ul
D E
∂F ∂F
Proof. First recall that gij = ∂u ,
i ∂uj
. By differentiating both sides respect to ul , we
get:
2 2
∂gij ∂ F ∂F ∂ F ∂F
= , + , .
∂ul ∂ul ∂ui ∂uj ∂ul ∂uj ∂ui
Using (7.14), we get:
* + * +
∂gij ∂F ∂F ∂F ∂F
(*) = Γkli + hli ν , + Γklj + hlj ν ,
∂ul ∂uk ∂uj ∂uk ∂ui
| {z } | {z }
∂2 F ∂2 F
∂ul ∂ui ∂ul ∂uj
7.3.2. The Proof of Theorema Egregium. The key ingredient of Gauss’s Theorema
Egregium is the following Gauss-Codazzi’s equations, which hold for any regular hyper-
surface in any dimension. When the dimension is two, the RHS of the Gauss’s equation
is similar to det[h] while the LHS depends only g. This gives Theorema Egregium as a
direct consequence of the Gauss’s equation.
Proof. The key step of the proof is to start with the fact that:
∂3F ∂3F
=
∂ui ∂uj ∂uk ∂uj ∂ui ∂uk
for any i, j, k, and then rewrite both sides in terms of the tangent-normal basis of Rn+1 .
Gauss’s equation follows from equating the tangent coefficients, and Codazzi’s equation
is obtained by equating the normal coefficient.
By (7.14), we have:
∂2F ∂F
= Γljk + hjk ν.
∂uj ∂uk ∂ul
Differentiating both sides with respect to ui , we get:
∂3F
∂ ∂F
= Γljk + hjk ν
∂ui ∂uj ∂uk ∂ui ∂ul
∂Γljk ∂F ∂2F ∂hjk ∂ν
= + Γljk + ν + hjk
∂ui ∂ul ∂ui ∂ul ∂ui ∂ui
∂Γljk ∂F
∂F ∂h jk ∂F
= + Γljk Γqil + hil ν + ν − hjk hqi
∂ui ∂ul ∂uq ∂ui ∂uq
q
∂Γjk ∂F
∂F ∂hjk ∂F
= +Γljk Γqil + + Γljk hil ν − hjk hql gil
∂ui ∂uq ∂uq ∂ui ∂uq
| {z }
l7→q
∂Γqjk
!
∂F ∂hjk
= + Γljk Γqil ql
− g hjk hil + + Γljk hil ν
∂ui ∂uq ∂ui
From (7.15) we also know that the Christoffel symbols depend only on the first funda-
mental form g but not on h. For simplicity, we denote:
q ∂Γqjk ∂Γqik
Rijk := − + Γljk Γqil − Γlik Γqjl .
∂ui ∂uj
q
The lower and upper indices for Rijk are chosen so as to preserve their positions in the
q
RHS expression (q being upper, and i, j, k being lower). We will see later that Rijk ’s are
the components of the Riemann curvature tensor.
We are now in a position to give a proof of Gauss’s Theorema Egregium as a direct
consequence of the Gauss’s equation.
7.3. Theorema Egregium 205
Proof. Consider the Gauss’s equation, which asserts that for any i, j, k and q:
q
Rijk = g ql (hjk hli − hik hlj ) .
Multiplying both sides by gpq , and summing up all q’s, we get:
q
gpq Rijk = gpq g ql (hjk hli − hik hlj ) = hjk hpi − hik hpj .
The above result is true for any i, j, k and p. In particular, when (i, j, k, p) = (1, 2, 2, 1),
we get:
q
g1q R122 = h22 h11 − h212 = det[h].
q
This shows det[h] depends only on g since R122 does so.
Finally, recall that the Gauss curvature is given by:
det[h]
K= .
det[g]
Therefore, we have completed the proof that K depends only g.
Remark 7.32. It is important to note that Theorem 7.31 holds for surfaces (i.e. dim Σ =
2).
q
The long Rijk -term can be interpreted in a nicer way using covariant derivatives. For
simplicity, we denote
∂F
∂i :=
∂ui
∇i := ∇∂i = ∇ ∂F
∂ui
∂Γljk
= ∂l + Γljk ∇i ∂l
∂ui
∂Γqjk
= ∂q + Γljk Γqil ∂q
∂ui
Similarly, we also have:
∂Γqik
∇j (∇i ∂k ) = ∂q + Γlik Γqjl ∂q .
∂uj
q
Hence, the term Rijk is the commutator of ∇i and ∇j :
[∇i , ∇j ]∂k = ∇i (∇j ∂k ) − ∇j (∇i ∂k )
∂Γqjk
!
∂Γqik
= − + Γljk Γqil − Γlik Γqjl ∂q
∂ui ∂uj
q
= Rijk ∂q .
In summary, we see that once gij is fixed, the covariant derivative ∇ and hence the
q
curvature term Rijk are uniquely determined. It motivates the idea that if we can
206 7. Geometry of Euclidean Hypersurfaces
declare the gij ’s on an abstract manifold M , then we can define its curvatures using the
prescribed gij even if M is not a submanifold of an Euclidean space. This motivates the
development of intrinsic geometry, a branch of geometry that plays no reference to the
ambient space but only the manifold itself. This branch is called Riemannian Geometry,
which is what this course is about!
Part 3
Riemannian Geometry
Chapter 8
Riemannian Manifolds
Johannes Kepler
209
210 8. Riemannian Manifolds
Example 8.4. Any regular hypersurface surface Σn in Rn+1 is a Riemannian manifold
with Riemannian metric g given by the first fundamental form. The symmetry and strict
positivity conditions are inherited from the flat metric on Rn+1 .
For instance, the round sphere S2 with radius R has a Riemannian metric given by:
g = R2 dϕ2 + R2 sin2 ϕ dθ2 .
Example 8.5. Given any smooth immersion Φ : Σ → M between two C ∞ smooths, and
suppose M is a Riemannian manifold with metric g. Then, Σ has an induced Riemannian
metric given by ge := Φ∗ g. The symmetry condition holds trivially. To verify the strict
positivity condition, we consider X ∈ Tp Σ, then
ge(X, X) = g(Φ∗ X, Φ∗ X) ≥ 0,
and by strict positivity of g, we have equality holds if and only if Φ∗ X = 0. Since Φ∗ is
injective (as Φ is an immersion), we must have X = 0.
In particular, any submanifold of a Riemannian manifold is itself a Riemannian
manifold with induced metric defined using the pull-back of the inclusion map.
Example 8.6 (Product Manifolds). Suppose (M, gM ) and (N, gN ) are two Riemannian
manifolds, then the product M × N is also a Riemannian manifold with metric given by:
∗ ∗
gM ⊕ gN := πM gM + πN gN
where πM : M × N → M and πN : M × N → N are projection maps. A tangent vector
X ∈ Tp (M × N ) can be expressed as (XM , XN ) ∈ Tp M ⊕ Tp N . The product metric acts
on tangent vectors by:
(gM ⊕ gN ) (XM , XN ), (YM , YN ) = gM (XM , YM ) + gN (XN , YN ).
Example 8.7 (Conformal Metrics). Given any Riemannian manifold (M, g) and a smooth
function f : M → R, one can define another Riemannian metric ge on M by conformal
rescaling:
ge = ef g.
Example 8.8 (Fubini-Study Metric on CPn ). On CPn , there is an important Riemannian
metric called the Fubini-Study metric, which is more convenient to be expressed using
complex coordinates. Parametrize CPn using standard local coordinates
h i
(j) (j) (j) (j) (j) (j)
F (j) z0 , · · · , zj−1 , zj+1 , · · · , zn(j) = z0 : · · · : zj−1 : 1 : zj+1 : · · · : zn(j) .
8.1. Riemannian Metrics 211
(j)
For simplicity, we let zj = 1. The Fubini-Study metric is defined to be:
Pn+1 (j) 2
2
∂ log k=1 z k
gFS := 2 Re dzp(j) ⊗ dz̄q(j) .
(j) (j)
∂zp ∂ z̄q
Given a Riemannian metric, one can then define the length of curve and volume of a
region of a manifold. Suppose γ(t) : [a, b] → M is a C 1 curve on a Riemannian manifold
(M, g), then we define the length of γ by (with respect to g) by:
Z bq
Lg (γ) := g γ 0 (t), γ 0 (t) dt
a
0 ∂
where γ (t) := γ∗ ∂t . Note that Lg (γ) depends on the metric g.
Given an open subset O of Riemannian manifold (M, g) such that O can be covered
by one single parametrization chart (u1 , · · · , un ), the volume of O (with respect to g) is
defined to be: Z q
Volg (O) = det[gij ] du1 ∧ · · · ∧ dun
O
∂
where gij = g( ∂u , ∂ ). The total volume of an orientable Riemannian manifold (M, g)
i ∂uj
is defined by:
XZ q
Volg (M ) := α ] du1 ∧ · · · ∧ dun
ρα det[gij α α
α Oα
8.1.1. Isometries. Recall that two smooth manifolds M and N are considered to
be the same in topological sense if there exists a diffeomorphism between them. Assume
further that they are Riemannian manifolds with metrics gM and gN . Heuristically, we
consider M and N to be geometrically the same if their Riemannian metrics are the “same”.
Precisely, take a diffeomorphism Φ : M → N , which sets up a one-one correspondence
between p ∈ M and Φ(p) ∈ N , and a tangent vector X ∈ Tp M corresponds to Φ∗ X ∈
TΦ(p) N . We consider gM and gN are “the same” if for any X, Y ∈ Tp M , we have
gM (X, Y ) = gN (Φ∗ X, Φ∗ Y ) meaning that inner product between X and Y is the same
as that between their correspondences Φ∗ X and Φ∗ Y . In this case, we then call Φ is an
isometry, and (M, gM ) and (N, gN ) are said to be isometric. Using the pull-back map, we
can rephrase these definitions in the following way:
Definition 8.9 (Isometries). Two Riemannian manifolds (M, gM ) and (N, gN ) are said
to be isometric if there exists a diffeomorphism Φ : M → N such that Φ∗ gN = gM . Such
a diffeomorphism Φ is called an isometry between (M, gM ) and (N, gN ).
Example 8.10. Consider the unit disk model D = {(x, y) ∈ R2 : x2 + y 2 < 1} of the
hyperbolic 2-space with the Poincaré’s metric introduced in Example 8.3:
4(dx2 + dy 2 )
g= .
1 − x2 − y 2
It is well-known that the hyperbolic plane can be equivalently modeled by the upper-half
plane U = {(u, v) : v > 0} with metric
du2 + dv 2
ge = .
v2
212 8. Riemannian Manifolds
Exercise 8.1. Complete the calculations in Example 8.10 to show that Φ is indeed
an isometry.
Without the ambient space, we now make sense of symmetries in an intrinsic way.
A C ∞ vector field V on (M, g) generates a 1-parameter subgroup of diffeomorphisms
Φt : M → M such that
∂Φt
(p) = V (Φt (p)), Φ0 = idM .
∂t
We then say g is symmetric along V if Φ∗t gΦt (p) = gp for any t ∈ R and p ∈ M .
Generally, such a diffeomorphism family Φt is hard to compute. Fortunately, one can
use the Lie derivative to check if g is symmetric along a given vector field without actually
computing Φt . The condition Φ∗t gΦt (p) = gp for any t ∈ R and p ∈ M is equivalent to
LV g = 0.
= ge(X, Y ).
∗
This shows π g = ge.
Example 8.12 (Flat Torus). A n-dimensional torus Tn can be regarded as the quotient
manifold Rn /Zn whose elements are equivalent classes [(x1 , · · · , xn )] under the relation
(x1 , · · · , xn ) ∼ (y1 , · · · , yn ) ⇐⇒ xi − yi ∈ Z for any i.
The quotient map π : Rn → Tn is then a covering map. Any Deck transformation
Φ : Rn → Rn is a diffeomorphism of Rn such that for any (x1 , · · · , xn ) ∈ Rn we have
[Φ(x1 , · · · , x1 )] = [(x1 , · · · , xn )]. In other words, there exists integers m1 , · · · , mn such
that
Φ(x1 , · · · , xn ) = (x1 , · · · , xn ) + (m1 , · · · , mn ).
These integers does not depend on xi ’s by continuity. Hence, the Deck transformation
group of the covering π : Rn → Tn is the set of translations by integer points.
It is clear that Φ∗ δ = δ where δ is the Euclidean metric of Rn . Hence by Proposition
8.11 (the converse part), there exists a metric g on Tn such that π ∗ g = δ.
Example 8.13 (Real Projective Space). Consider RPn constructed by identifying an-
tipodal points on Sn . The quotient map π : Sn → RPn is a covering map, and Deck
transformations Φ : Sn → Sn of this covering are either Φ = id or Φ = −id. Both are
isometries of the round metric ground of Sn , so it induces a Riemannian metric g such that
π ∗ g = ground .
214 8. Riemannian Manifolds
∂ ∂
Exercise 8.2. Let X = X i ∂u i
and Y = Y j ∂u j
. Show that ∇X Y has the following
local expression:
∂Y k
∂
(8.4) ∇X Y = X i + X i Y j Γkij .
∂ui ∂uk
Check also that (8.4) is independent of local coordinates, i.e. given another local
coordinates {vα } so that X = X e α ∂ , Y = Ye β ∂ , and geαβ = g( ∂ , ∂ ), verify
∂vα ∂vα ∂vα ∂vβ
that: !
eγ k
e α ∂Y + X
X eγ
e α Ye β Γ ∂
= X i ∂Y
+ X i j k
Y Γ ij
∂
αβ
∂vα ∂vγ ∂ui ∂uk
e γ in the new coordinate system are defined as
where the Christoffel symbols Γ αβ
e γ = 1 geγη ∂e
Γ
gβη
+
∂e
gαη
−
∂e
gαβ
.
αβ
2 ∂e
vα ∂e
vβ ∂evη
The operator ∇ described above is called the Levi-Civita connection (or Riemannian
connection) of (M, g). Note that the Christoffel symbols Γkij , and hence the Levi-Civita
connection, depend on g.
The term connection has a broader meaning. Any operator D : Γ∞ (T M )×Γ∞ (T M ) →
∞
Γ (T M ) satisfying all of (8.1)-(8.3) is called a connection on M .
8.2. Levi-Civita Connections 215
It is straight-forward to verify that Γkij and gij satisfy the following relations:
∂X i ∂Y j
∂ ∂gij i j
Z (g(X, Y )) = Z k gij X i Y j = Z k X Y + gij Y j + gij X i
.
∂uk ∂uk ∂uk ∂uk
∂
On the other hand, we have (for simplicity, denote ∂k := ∂uk and ∇k := ∇∂k ):
g(∇Z X, Y ) + g(X, ∇Z Y )
= g Z k ∇k (X i ∂i ), Y j ∂j + g X i ∂i , Z k ∇k (Y j ∂j )
i
k ∂X ∂ k i l ∂ j ∂
=g Z + Z X Γki ,Y
∂uk ∂ui ∂ul ∂uj
j
∂ ∂Y ∂ ∂
+ g Xi , Zk + Z k Y j Γlkj
∂ui ∂uk ∂uj ∂ul
∂X i j ∂Y j i
= Z k glj Γlki X i Y j + gil Γlkj X i Y j +gij Z k Y + gij Z k
X .
| {z } ∂uk ∂uk
∗
k
Proof. Let (ui ) be local coordinates of M and define γij by
∂ k ∂
D ∂ = γij .
∂ui ∂uj ∂uk
Then, the given two conditions of this proposition are equivalent to
k k
γij = γji
∂gij l l
= γik gjl + γjk gil
∂uk
for any i, j, k. The proof is identical to the one that shows (8.5)-(8.6) and (8.7)-(8.8) are
equivalent (see page 215). By cyclic permutation of indices, we can then show:
∂gjk ∂gik ∂gij l
+ − = 2γij gkl
∂ui ∂uj ∂uk
for any i, j, k. Multiplying g −1 on both sides, we get:
l 1 kl ∂gjk ∂gik ∂gij
γij = g + − ,
2 ∂ui ∂uj ∂uk
which is exactly the Christoffel symbols Γkij for the Levi-Civita connection of g. Since the
action of D is uniquely determined by its action on coordinate vectors, we must have
D = ∇.
Therefore, one can also define the Levi-Civita connection with respect to (M, g) as
the unique operator that satisfies all of (8.1)-(8.3) and (8.7)-(8.8).
8.2.1. Tensorial quantities. A linear operator T acting on vector fields and 1-forms
(X1 , · · · , Xr , ω1 , · · · , ωs ) ∈ Γ∞ (T M )r × Γ∞ (T ∗ M )s is said to be tensorial if one can
factor out a scalar function in each slot, i.e.
T (ϕ1 X1 , · · · , ϕr Xr , ψ1 ω1 , · · · , ψs ωs ) = ϕ1 · · · ϕr ψ1 · · · ψs T (X1 , · · · , Xr , ω1 , · · · , ωs )
for any scalar functions ϕi ’s and ψi ’s.
A Riemannian metric g is tensorial as it is required to be. For Euclidean hypersurfaces,
the second fundamental form h is tensorial since Df X ν = f DX ν. However, the operator
8.2. Levi-Civita Connections 217
= (Xf )Y + f DX Y − (Xf )Y − f D
eX Y
= f (D − D)(X,
e Y ).
Given any vector field X and 1-form α, the output ∇X α is a 1-form such that given
any vector field Y the product rule holds formally:
X (α(Y )) = (∇X α)(Y ) + α(∇X Y )
or equivalently,
(∇X α)(Y ) := X(α(Y )) − α(∇X Y ).
Locally, we have:
∂ ∂ ∂ ∂
(∇i duj ) = duj − duj ∇i
∂uk ∂ui ∂uk ∂uk
∂ j ∂
= δ − duj Γlik
∂ui k ∂ul
= 0 − Γlik δlj = −Γjik .
In other words, we have
∇i duj = −Γjik duk .
Then, given any vector fields X, Y1 , · · · , Ys and 1-forms α1 , · · · , αr , we define the
operator ∇X by:
∇X (Y1 ⊗ Y2 ) := (∇X Y1 ) ⊗ Y2 + Y1 ⊗ (∇X Y2 )
and more generally,
r
X
∇X (⊗ri=1 ωi ) ⊗ (⊗sj=1 Yj ) =
ω1 ⊗ · · · ⊗ ∇X ωi ⊗ · · · ⊗ ωr ⊗ Y1 ⊗ · · · ⊗ Ys
i=1
s
X
+ ω1 ⊗ · · · ⊗ ωr ⊗ Y1 ⊗ · · · ⊗ ∇X Yj ⊗ · · · ⊗ Ys .
j=1
∂Tij i
= du ⊗ duj + Tij (∇k dui ) ⊗ duj + Tij dui ⊗ (∇k duj )
∂uk
∂Tij i
= du ⊗ duj − Tij Γikl dul ⊗ duj − Tij Γjkl dui ⊗ dul
∂uk
∂Tij
= − Tlj Γlki − Til Γlkj dui ⊗ duj .
∂uk
We denote ∇k Tij to be the local expression such that
∇k Tij dui ⊗ duj =: (∇k Tij ) dui ⊗ duj ,
∂Y i ∂
i ∂ ∂
∇j Y = + Y i ∇j
∂ui ∂uj ∂ui ∂ui
∂Y i ∂ ∂
= + Y i Γkji
∂uj ∂ui ∂uk
∂Y i
k i ∂
= + Y Γjk .
∂uj ∂ui
∂Y i
∇j Y i = + Y k Γijk .
∂uj
···js ∂ ∂
T = Tij11···i dui1 ⊗ · · · ⊗ duir ⊗ ⊗ ··· ⊗ ,
r
∂uj1 ∂ujs
···js
we define ∇k Tij11···i r
to be the local expression such that
···js ∂ ∂
∇k Tij11···i dui1 ⊗ · · · ⊗ duir ⊗ ⊗ ··· ⊗
r
∂uj1 ∂ujs
···js
∂ ∂
= ∇k Tij11···i dui1 ⊗ · · · ⊗ duir ⊗ ⊗ ··· ⊗ .
r
∂uj1 ∂ujs
Exercise 8.9. Show that the Codazzi equation (Theorem 7.30) is equivalent to
∇i hjk = ∇j hik
for any i, j, k.
Since ∇X is tensorial in the X-slot, given a (r, s)-tensor T , one can then view X as
an input and define a new (r + 1, s)-tensor ∇T as
∂
Locally, given a vector field Y = Y i ∂u i
, we have:
∂ ∂
∇j Y = (∇j Y i ) =⇒ ∇Y := duj ⊗ (∇j Y ) = (∇j Y i ) duj ⊗ .
∂ui ∂ui
220 8. Riemannian Manifolds
It is important to note that ∇i and ∇i are different! The former was discussed above,
but we define
∇i := g ij ∇j .
For instance, when acting on scalar functions, we have
∂f ∂f
∇i f = whereas ∇i f = g ij ∇j f = g ij .
∂ui ∂uj
∂
For a vector field Y = Y i ∂u i
, we have:
∂Y i
∇j Y i = g jk ∇k Y i = g jk + Y l Γikl .
∂uk
∂
While we can define ∇Y as a (1, 1)-tensor (∇i Y j ) dui ⊗ ∂u j
, it also makes sense to regard
it as a (0, 2)-tensor:
∂ ∂ ∂ ∂
(∇i Y j ) ⊗ = (g ik ∇k Y j ) ⊗ .
∂ui ∂uj ∂ui ∂uj
Very often, it is not difficult to judge whether ∇Y means a (1, 1)-tensor or a (0, 2)-tensor
according to the context. Similar notations apply to other higher-rank tensors.
However, when acting on scalar functions f , it is a convention to regard ∇f as a
vector field but not a 1-form, that is:
∂ ∂f ∂
∇f := (∇i f ) = g ij
∂ui ∂uj ∂ui
∂f
instead of (∇i f ) dui = ∂u i
dui . We have this convention because we already have another
∂f
symbol to denote ∂u i
dui , namely the exterior derivative df . The vector field ∇f is called
8.2. Levi-Civita Connections 221
the gradient of f with respect to g. When g is the Euclidean metric on Rn , the gradient
∇f is the standard gradient in multivariable calculus.
Using the Levi-Civita connection, one can also define divergence and Laplacian
∂
operators. Given a vector field X = X i ∂u i
on a Riemannian manifold (M, g), we define
its divergence with respect to g by
divg X := ∇i X i .
The Laplacian operator ∆ acting on functions is the divergence of the gradient with
respect to g:
∆g f := divg (∇f ) = ∇i ∇i f.
Since ∇i = g ij ∇j and g is constant under ∇, we have
2
ij ij ij ∂ f k ∂f
∆g f = ∇i (g ∇j f ) = g ∇i ∇j f = g − Γij .
∂ui ∂uj ∂uk
8.2.3. Contraction of tensors. Given two tensors of different types, one can “cook
them up” to form new tensors. For instance, consider the first fundamental form gij (and
its inverse g ij ), and the second fundamental form hij of a regular Euclidean hypersurface.
They are both (2, 0)-tensors. We can define “cook up” two new tensors by summing up
indices:
g ij hij or g ik hkj .
They are different tensors. The first one g ij hij sums over both i and j, so there is no
free component and it is a scalar function (this function is the mean curvature). For the
second one g ik hkj , we sum them up over k leaving i and j to be free. It gives a new
tensor denoted by, say, A with one lower component and one upper component (i.e. a
∂
(1, 1)-tensor), that if we input ∂u j
into A, it will output
∂ ∂ ∂
A = g ik hkj , or equivalently, A = g ik hkj duj ⊗ .
∂uj ∂ui ∂ui
In this case, the operator A is simply the shape operator discussed in Section 7.2. We
often denote the components of A by hij := g ik hkj .
For more complicated tensors, we can “cook them up” in similar ways. For example,
k
given S = Sij dui ⊗ duj ⊗ ∂u∂ k and T = Tji duj ⊗ ∂u
∂
i
, we can form many combinations:
k j k j
αi := Sij Tk (1, 0)-tensor α = Sij Tk dui
It is interesting to note that if each summations is over exactly one upper index and
one lower index, then the new tensor produced is independent of local coordinates. Take
222 8. Riemannian Manifolds
∂
= (Sjk Tqk ) duj ⊗ duq + Sjk Tqk (∇i duj ) ⊗ duq + Sjk Tqk duj ⊗ (∇i duq )
∂ui
!
∂Sjk k ∂Tqk
= T + Sjk duj ⊗ duq − Sjk Tqk Γjip dup ⊗ duq − Sjk Tqk duj ⊗ (Γqip dup )
∂ui q ∂ui
!
∂Sjk k ∂Tqk
= T + Sjk duj ⊗ duq − Spk Tqk Γpij duj ⊗ duq − Sjk Tpk Γpiq duj ⊗ duq .
∂ui q ∂ui
The last step follows from relabelling indices of the last two terms. By taking out the
common factor duj ⊗ duq , we have proved
∂Sjk k ∂Tqk
∇i (Sjk Tqk ) = Tq + Sjk − Spk Tqk Γpij − Sjk Tpk Γpiq .
∂ui ∂ui
For the RHS, we recall that:
∂Sjk
∇i Sjk = − Γpij Spk − Γpik Sjp ,
∂ui
∂Tqk
∇i Tqk = − Γpiq Tpk + Γkip Tqp .
∂ui
Combining these results, we can easily get that
(∇i Sjk )Tqk + Sjk (∇i Tqk ) = ∇i (Sjk Tqk )
by cancellations and relabelling of indices.
A similar product rule holds for all other legitimate contractions (i.e. one “up” paired
with one “down”). The proof is similar to the above special case although it is quite
tedious. We omit the proof here.
8.2. Levi-Civita Connections 223
In particular, as we have ∇i gjk = 0 for any i, j, k, one can apply the product rule to
see that gij can be treated as a constant when we differentiate it by ∇. For example:
∇i (gjk Tlk ) = gjk ∇i Tlk .
Recall also that gik g kj = δij . Applying the product rule one can get
∇p (gik g kj ) = 0 =⇒ gik ∇p g kj = 0.
Taking g −1 on both sides, we can get ∇p g kj = 0 as well.
Chapter 9
The Mother
225
226 9. Parallel Transport and Geodesics
dγ i ∂
γ 0 (t) =
dt ∂ui
∂
∇γ 0 (t) V (t) = ∇ dγ i ∂ V j (t)
dt ∂ui ∂uj
i j
dγ ∂V ∂ j k ∂
= + V Γij
dt ∂ui ∂uj ∂uk
dV j ∂ dγ i ∂
= + V j Γkij .
dt ∂uj dt ∂uk
Therefore, the parallel transport equation ∇γ 0 (t) V (t) = 0 is equivalent to
dV k dγ i
+ V j Γkij = 0 for any k.
dt dt
Notably, the equation depends on γ 0 (t) only but not on γ(t) itself, so it also makes sense
of parallel transporting V0 along a vector field X. It simply means parallel transport of
V0 along the integral curve γ(t) such that γ 0 (t) = X(γ(t)).
Even though the parallel transport equation is a first-order ODE whose existence
and uniqueness of solutions are guaranteed, it is often impossible to solve it explicitly
(and often not necessary to). Nonetheless, there are many remarkable uses of parallel
transports. When we have an orthonormal basis {ei }ni=1 of Tp M at a fixed point p ∈
M , one can naturally extend it along a curve to become a moving orthonormal basis
{ei (t)}ni=1 along that curve. The reason is that parallel transport preserves the angle
between vectors.
Given a curve γ(t) on a Riemannian manifold (M, g), and let V (t) be the parallel
transport of V0 ∈ Tp M along γ, and W (t) be that of W0 ∈ Tp M . Then one can check
easily that
d
g V (t), W (t) = g ∇γ 0 V, W + g(V, ∇γ 0 W = 0
dt
by the parallel transport equations ∇γ 0 V = ∇γ 0 W = 0. Therefore, we have
g V (t), W (t) = g V0 , W0
for any t. In particular, if V0 and W0 are orthogonal, then so are V (t) and W (t) for any t.
If we take W0 = V0 , by uniqueness theorem p of ODE we have W (t) ≡ V (t). The above
result shows the length of V (t), given by g(V (t), V (t)), is also a constant. Hence, the
parallel transport of an orthonormal basis remains to be orthonormal along the curve.
Exercise 9.1. Let (M, g) be a connected complete Riemannian manifold, and p and
q be two distinct points on M . Show that Holp (M, g) and Holq (M, g) are related by
conjugations (hence are isomorphic).
When (M, g) is the flat Euclidean space Rn , parallel transport is simply translations.
Any vector will end up being the same vector after transporting back to its based point.
Therefore, Holp (Rn , δ) is the trivial group for any p.
For the round sphere S2 in R3 , we pick two points P , Q on the equator, and mark
N to be the north pole. Construct a piecewise great circle path P → N → Q → P , then
one can see that what parallel transport along this path does is a rotation on vectors in
TP S2 by an angle depending on the distance between P and Q. For instance, when P
and Q are antipodal, the parallel transport map is rotation by π. When P = Q, then the
parallel transport map is simply the identity map. By varying the position of Q, one can
obtain all possible angles from 0 to 2π. To explain this rigorously, one way is to solve the
parallel transport equation in Definition 9.1. There is a more elegant to explain this after
we learn about geodesics.
It is remarkable that the holonomy group reveals a lot about the topological structure
about a Riemannian manifold, and is a extremely useful tool for classification problems
of manifolds. There is a famous theorem due to Ambrose-Singer that the Lie algebra of
Holp (M, g) is related to how curved (M, g) is around p.
Another beautiful theorem which demonstrates the usefulness of parallel transport is
the following one due to de Rham. It is widely used in Ricci flow to classify the topology
of certain class of manifolds by decomposing them into lower dimensional ones.
Theorem 9.3 (de Rham Splitting Theorem - local version). Suppose the tangent bundle
T M of a Riemannian manifold (M, g) can be decomposed orthogonally into T M = E1 ⊕E2
such that each of Ei ’s is invariant under parallel transport, i.e. whenever V ∈ Ei , any
parallel transport of V stays in Ei . Then, M is locally a product manifold (N1 , h) × (N2 , k)
such that T Ni = Ei , and g is a locally product metric π1∗ h + π2∗ k.
Proof.
∂The proof begins byconstructing a local coordinate system {xi , yα } such that
∂
span ∂x i
= E 1 and span ∂yα = E2 . This is done by Frobenius’ Theorem which
asserts that such that such local coordinate system would exist if each Ei is closed under
the Lie brackets, i.e. whenever X, Y ∈ Ei , we have [X, Y ] ∈ Ei .
Let’s prove E1 is closed under the Lie brackets (and the proof for E2 is exactly the
same). The key idea is to use (8.7) that:
[X, Y ] = ∇X Y − ∇Y X.
First pick an orthonormal basis {ei , eα } at a fixed point p ∈ M , such that span ei = E1
and span eα = E2 . Extend them locally around p using parallel transport, then one
228 9. Parallel Transport and Geodesics
would have ∇ei = ∇eα = 0 for any i and α. Now suppose X, Y ∈ E1 , we want to show
∇X Y ∈ E1 . Since g(Y, eα ) = 0 and ∇X eα = 0, we have by (8.8):
0 = X g(Y, eα ) = g(∇X Y, eα ).
This shows ∇X Y ⊥ eα for any α, so ∇X Y ∈ E1 . The same argument shows ∇Y X ∈ E1 ,
and so does [X, Y ].
Frobenius’
∂ ∂ that there exists a local coordinate system {xi , yα } with
Theorem asserts
span ∂x i
= E1 and span ∂yα = E2 . Next, we argue that the metric g is locally
expressed as:
g = gij dxi ⊗ dxj + gαβ dy α ⊗ dy β ,
and that gij depends only on {xi }, and gαβ depends only on {yα }. To show this, we
j β
argue that Christoffel symbols of type Γα i
ij , Γiα , Γiα , and Γαβ are all zero.
∂ ∂
Consider V0 := ∂xi ∈ Tp M , and let V (t) be its parallel transport along ∂xj . Since
V (t) ∈ E1 , it can be locally expressed as V (t) = V k
(t) ∂x∂ k .
We then have
∂V k ∂
k l ∂ α ∂
0 = ∇j V (t) = +V Γjk + Γjk .
∂xj ∂xk ∂xl ∂yα
By equating coefficients, we get V k Γα α
jk = 0. In particular, it implies Γik = 0 at p
∂
since V (0) = ∂x i
at p. Apply the same argument at other points covered by the local
coordinates {xi , yα }, we have proved Christoffel symbols of type Γα
ik all vanish. Using
a similar argument by parallel transporting V0 along ∂y∂α instead, one can also show
Γβαi = 0. The other combinations Γjiα and Γiαβ follow from symmetry argument.
Finally, we conclude that
∂
gij = Γkαi gkj + Γkαj gki = 0,
∂yα
so gij depends only on {xi }. Similarly, gαβ depends only on {yα }. It completes our
proof.
Remark 9.4. There is a global version of de Rham splitting theorem. If one further
assumes that M is simply-connected, then the splitting would be in fact global. Generally
if (M, g) satisfies the hypothesis of the de Rham splitting theorem, one can consider
its universal cover (Mf, π ∗ g) which is simply-connected. Applying the theorem one can
f, π ∗ g) splits isometrically as a product manifold.
assert that (M
Now parallel transport V along an arbitrarily given curve γ starting from P , then we
have:
∇γ 0 Tij V i − λV j = (∇γ 0 Tij )V i + Tij ∇γ 0 V i − λ∇γ 0 V j = 0
Proposition 9.5. Let γs (t) : (−ε, ε) × [a, b] → M be a 1-parameter family of curves with
γs (a) = p and γs (b) = q for any s ∈ (−ε, ε). Here s is the parameter of the family, and t
is the parameter of each curve γs . Then, the first variation of the arc-length is given by:
Z b 0
d ∂γ γ0 (t)
(9.1) L(γs ) = − g , ∇γ00 (t) dt.
ds s=0 a ∂s s=0 |γ00 (t)|
Therefore, if γ0 minimizes the length among all variations γs , it is necessarily that
0
γ0 (t)
(9.2) ∇γ0 (t)
0 = 0 for any t ∈ [a, b].
|γ00 (t)|
Exercise 9.3. Compute the first variation of the energy functional E(γs ) of a 1-
parameter family of curves γs (t) : (−ε, ε) × [a, b] → M with γs (a) = p and γs (b) = q
for any s. The energy functional is given by
1 b
Z
∂γs ∂γs
E(γs ) := g , dt.
2 a ∂t ∂t
Note that it is different from arc-lengths: there is not square root in the integrand.
Furthermore, show that if γ0 minimizes E(γs ), then one has ∇γ00 (t) γ00 (t) = 0 for
any t ∈ [a, b].
Since every (regular) curve can be parametrized by constant speed, we can assume
without loss of generality that |γ00 (t)| ≡ C, so that (9.2) can be rewritten as
(9.3) ∇γ00 γ00 = 0.
We call (9.3) the geodesic equation and any constant speed curve γ0 (t) that satisfies (9.3)
is called a geodesic. For simplicity, let’s assume from now on that every geodesic has a
constant speed. We can express (9.3) using local coordinates. Suppose under a local
coordinate system F (u1 , · · · , un ), the coordinate representation of γ0 is given by:
F −1 ◦ γ0 (t) = (γ 1 (t), · · · , γ n (t)).
Then γ00 (t) is given by
dγ i ∂
γ00 := ,
dt ∂ui
so we have
dγ j ∂
∇γ00 γ00 = ∇γ00
dt ∂uj
d2 γ j ∂ dγ j ∂
= 2
+ ∇ dγ i ∂
dt ∂uj dt dt ∂ui ∂uj
d2 γ j ∂ dγ i dγ j k ∂
= + Γ
dt2 ∂uj dt dt ij ∂uk
2 k
dγ i dγ j k
d γ ∂
= 2
+ Γij
dt dt dt ∂uk
Hence, the geodesic equation (9.3) is locally a second-order ODE (assuming γ0 is arc-
length parametrized):
d2 γ k dγ i dγ j k
(9.4) + Γ = 0 for any k.
dt 2 dt dt ij
Example 9.6. Consider the round sphere S2 in R3 which, under spherical coordinates,
has its Riemannian metric given by
g = sin2 θ dϕ2 + dθ2 .
By direct computations using (7.15), we get:
Γθϕϕ = − sin θ cos θ
Γϕ ϕ
ϕθ = Γϕθ = cot θ
Write γ(t) locally as (γ ϕ (t), γ θ (t)), then the geodesic equations are given by
d2 γ ϕ dγ ϕ dγ θ ϕ dγ θ dγ ϕ ϕ
+ Γ + Γ =0
dt2 dt dt ϕθ dt dt θϕ
d2 γ θ dγ ϕ dγ ϕ θ
+ Γ = 0,
dt2 dt dt ϕϕ
that is
d2 γ ϕ dγ ϕ dγ θ
2
+2 cot θ = 0
dt dt dt
ϕ 2
d2 γ θ dγ
2
− sin θ cos θ = 0.
dt dt
Clearly, the path with (γ θ (t), γ ϕ (t)) = (t, c), where c is a constant, is a solution to the
system. This path is a great circle.
Exercise 9.5. Consider the surface of revolution F (u, θ) = (x(u) cos θ, x(u) sin θ, z(u))
given by a profile curve (x(u), 0, z(u)) on the xz-plane. Show that the profile curve
itself is a geodesic of the surface.
Exercise 9.6. Consider the hyperbolic space with the upper-half plane model, i.e.
dx2 + dy 2
H2 = {(x, y) ∈ R2 : y > 0}, g = .
y2
Show that straight lines normal to the x-axis, as well as semi-circles intersecting the
x-axis orthogonally, are geodesics of the hyperbolic space.
Since (9.4) is a second-order ODE system, it has local existence and uniqueness when
given both initial position and velocity. That is, given p ∈ M and V ∈ Tp M , there exists a
unique (constant speed) geodesic γV : (−ε, ε) → M such that γV (0) = p and γV0 (0) = V .
Let c > 0, the geodesic γcV (starting from p) is one that the speed is c |V |, and
so it travels c times faster than γV does. One should then expect that γcV (t) = γV (ct)
provided ct is in the domain of γV . It can be easily shown to be true using uniqueness
theorem ODE, since γV (ct) t=0 = γV (0) = p and
d
γV (ct) = cγV0 (ct) t=0
= cγV0 (0) = cV.
dt t=0
Therefore, γV (ct) is a curve initiating from p with initial velocity cV . It can be shown
using the chain rule that γV (ct) also satisfies (9.4). Hence, by uniqueness theorem of
ODE, we must have γV (ct) = γcV (t).
Note that a geodesic γ(t) may not be defined globally on (−∞, ∞). Easy counter-
examples are straight-lines ax + by = 0 on R2 \{0}. If every geodesic γ passing through p
on a Riemannian manifold (M, g) can be extended so that its domain becomes (−∞, ∞),
then we say (M, g) is geodesically complete at p. If (M, g) is geodesically complete is every
point p ∈ M , then we say (M, g) is geodesically complete.
9.2. Geodesic Equations 233
Exercise 9.7. Check that d is a metric on M (in the sense of a metric space).
On a metric space, we can talk about Cauchy completeness, meaning that every
Cauchy sequence in M with respect to d converges to a limit in M . Interestingly, the two
notions of completeness are equivalent! It thanks to:
Theorem 9.7 (Hopf-Rinow). Let (M, g) be a connected Riemannian manifold, and let
d : M × M → R be the distance function induced by g. Then the following are equivalent:
(1) There exists p ∈ M such that (M, g) is geodesically complete at p
(2) (M, g) is geodesically complete
(3) (M, d) is Cauchy complete
We omit the proof in this note as it “tastes” differently from other parts of the course.
Interested readers may consult any standard reference of Riemannian geometry to learn
about the proof.
1Here metric means the distance function of a metric space, not a Riemannian metric on a manifold!
234 9. Parallel Transport and Geodesics
Definition 9.8 (Exponential Map). Let (M, g) be a complete Riemannian manifold. Fix
a point p ∈ M and given any tangent vector V ∈ Tp M , we consider the unique geodesic
γV with initial conditions γV (0) = p and γV0 (0) = V . We define expp : Tp M → M by
expp (V ) := γV (1).
The map expp is called the exponential map at p.
Remark 9.9. Standard ODE theory shows expp is a smooth map. For detail, see e.g. p.74
of John M. Lee’s book.
Proof. The key observation is the rescaling property γcV (t) = γV (ct), which implies
expp (cV ) = γcV (1) = γV (c), and hence
d d
(expp )(0 + tV ) = γV (t) = γV0 (t) = V.
dt t=0 dt t=0
In other words, (expp )∗0 (V ) = V .
Clearly, (expp )∗0 is invertible, and by inverse function theorem we can deduce that
expp is locally a diffeomorphism near 0 ∈ Tp M :
Corollary 9.11. There exists an open ball B(0, ε) on Tp M such that expp B(0,ε)
is a
diffeomorphism onto its image.
To sum up, expp (V ) is well-defined for all V ∈ Tp M provided that (M, g) is complete.
However, it may fail to be a diffeomorphism if the length of V is too large. The maximum
possible length is called:
Example 9.13. Injectivity radii may not be easy to be computed, but we have the
following intuitive examples:
• For the round sphere S2 with radius R, the injectivity radius at every point is given
by πR.
• For the flat Euclidean space, the injectivity radius at every point is +∞.
• For a round torus obtained by rotating a circle with radius r, the injectivity radius
at every point is πr.
Proof. The key idea is to slightly modify the map expp . By Gram-Schmidt’s orthogonal-
ization, one can take a basis {êi }ni=1 for Tp M which are orthonormal with respect to g,
i.e. g(êi , êj ) = δij . Then we define an isomorphism E : Rn → Tp M by:
E(u1 , · · · , un ) := ui êi .
Next we define the parametrization G : E −1 (B(0, ε)) → M by G := expp ◦E. Here ε > 0
is sufficiently small so that expp B(0,ε) is a diffeomorphism onto its image.
Next we claim such G satisfies all three required conditions. To prove (1), we
compute that:
∂ ∂G ∂ ∂
(p) := = G∗ (0) = (expp )∗ ◦ E∗ (0) .
∂ui ∂ui 0 ∂ui ∂ui
| {z } | {z }
∈Tp M ∈T0 Rn
∂
From the definition of E, we have E∗ ( ∂u i
) = êi , and also at p, we have:
(expp )∗0 (êi ) = êi
∂
according to Lemma 9.10. Now we have proved ∂u i
(p) = êi . Hence, (1) follows directly
from the fact that {êi } is an orthonormal basis with respect to g:
∂ ∂
g (p), (p) = g(êi , êj ) = δij .
∂ui ∂uj
Next we claim that (2) is an immediate consequence of the geodesic equation (9.4).
Consider the curve γ(t) = expp (t(êi + êj )), which is a geodesic passing through p. Then,
the local expression of γ(t) is given by:
G−1 ◦ γ(t) = (γ 1 (t), · · · , γ n (t))
236 9. Parallel Transport and Geodesics
where (
t if k = i or j
γ k (t) = .
0 otherwise
By (9.4), we have:
d2 γ k i
k dγ dγ
j
2
+Γij = 0.
|dt
{z } |dt{zdt}
=0 =1
This shows Γkij = 0 along γ(t) (see footnote2), and in particular Γkij = 0 at p = γ(0).
Finally, (3) is an immediate consequence of (8.6).
The geodesic normal coordinates are one of the “gifts” to Riemannian geometry. In
the later part of the course when we discuss geometric flows, we will see that it simplifies
many tedious tensor computations. It is important to note that different points give
rise to different geodesic normal coordinate systems! Note also that a geodesic normal
coordinate system satisfies the three properties in Proposition 9.14 at one point p only.
It is not always possible to pick a local coordinate system such that gij = δij on the whole
chart, unless the Riemannian manifold is locally flat.
Here is one demonstration of how useful geodesic normal coordinates are. Suppose
M is a Riemannian manifold with a smooth family of Riemannian metrics g(t), t ∈ [0, T ),
such that g(t) evolves in the direction of a family of symmetric 2-tensor v(t), i.e.
∂
g(t) = v(t).
∂t
∂
In terms of local components, we may write ∂t gij (t) = vij (t) where
g(t) = gij (t) dui ⊗ duj , v(t) = vij (t) dui ⊗ duj .
We want to derive the rate of change of the Christoffel symbols Γkij (t)’s, which change
over time.
∂ k
First we fix a time and a point (t0 , p). Note that ∂t Γij is tensorial since the difference
k k k
Γij (t) − Γij (t0 ) is tensorial (even though Γij (t) itself isn’t). Therefore, if one can express
∂ k
∂t Γij at (t0 , p) as another tensorial quantity using one particular local coordinate system,
then this expression holds true under all other local coordinate systems.
Let’s choose the geodesic normal coordinates {ui } at p with respect to the metric
g(t0 ). Then, we have
∂gij
gij (t0 , p) = δij , Γkij (t0 , p), (t0 , p) = 0.
∂uk
Recall that from (7.15) we have
k 1 kl ∂gjl ∂gil ∂gij
Γij = g + − .
2 ∂ui ∂uj ∂ul
Taking time derivatives, we have:
∂ k 1 ∂ kl ∂gjl ∂gil ∂gij 1 kl ∂ ∂gjl ∂ ∂gil ∂ ∂gij
Γ = g + − + g + − .
∂t ij 2 ∂t ∂ui ∂uj ∂ul 2 ∂ui ∂t ∂uj ∂t ∂ul ∂t
2Note that Γk here is in fact Γk (γ(t)), so one can only claim Γk = 0 along γ(t) but not at other points. Also, γ(t)
ij ij ij
itself depends on the choice of i and j. This argument does not imply there is a geodesic on which all Christoffel’s symbols
vanish.
9.3. Exponential Map 237
∂g
However, when evaluating at (t0 , p), the space derivatives ∂ujli ’s all become zero, so we
can ignore the first term above and obtain
∂Γkij
1 ∂vjl ∂vil ∂vij
(t0 , p) = g kl + − .
∂t 2 ∂ui ∂uj ∂ul (t0 ,p)
∂v
Note that ∂ujli ’s are not tensorial, but at (t0 , p) all Christoffel symbols are zero, so by
Exercise 8.7 we have
∂vjl
= ∇i vjl at (t0 , p),
∂ui
and so
∂Γkij 1
= g kl (∇i vjl + ∇j vil − ∇l vij ) at (t0 , p).
∂t 2
Note that now both sides are tensorial, so given that the above equation holds for one
particular coordinate system, it holds for all other coordinate systems. Say {yα }, we will
still have
∂Γγαβ 1
= g γη ∇α vβη + ∇β vαη − ∇η vαβ at (t0 , p).
∂t 2
Furthermore, even though this derivative expression holds at (t0 , p) only, we can repeat
the same argument using geodesic normal coordinates with respect g(t) at other time
and at other points, so that we can conclude under any local coordinates on M , we have:
∂Γkij 1
= g kl (∇i vjl + ∇j vil − ∇l vij ) .
∂t 2
Curvatures of Riemannian
Manifolds
239
240 10. Curvatures of Riemannian Manifolds
for the (3, 1)-tensor. However, even though it indeed gives T122 = ∇1 ∇2 ∂2 − ∇2 ∇1 ∂2 ,
one can easily verify that such T is not tensorial. We modify such an T a bit and define
the following very important tensors in Riemannian geometry:
If the tensor type is clear from the context, we may simply call it the Riemann curvature
tensor and simply denote it by Rm. Alternative notations include R, R, etc.
∂ ∂
When X = ∂ui and Y = ∂uj , we have [X, Y ] = 0, so that it still gives
It is clear from the definition that Rm(3,1) (X, Y )Z = −Rm(3,1) (Y, X)Z. That explains
why we intentionally write its input vectors by (X, Y )Z instead of (X, Y, Z), so as to
emphasize that X and Y are alternating.
We express the local components of Rm(3,1) by Rijk
l
, so that
∂
Rm(3,1) = Rijk
l
dui ⊗ duj ⊗ duk ⊗
∂ul
where
∂ ∂ ∂ ∂
l
Rijk = Rm(3,1) ,
∂ul ∂ui ∂uj ∂uk
= ∇i ∇j ∂k − ∇j ∇i ∂k − ∇[∂i ,∂j ] ∂k
!
∂Γljk ∂Γlik p l p l ∂
= − + Γjk Γip − Γik Γjp .
∂ui ∂uj ∂ul
∂Y k 1 ∂2Y k
Y k (u1 , 0) = Y k (0, 0) + (u1 − 0) + (u1 − 0)2 + O(u31 ).
∂u1 (0,0) 2 ∂u21 (0,0)
Using the parallel transport equation, both the first and second derivatives of Y k can be
expressed in terms of Christoffel’s symbols.
∂Y k
= −Y j Γk1j
∂u1
∂2Y k ∂Y j k j
∂Γk1j
= − Γ − Y
∂u21 ∂u1 1j ∂u1
∂Γk1j
= Y l Γj1l Γk1j − Y j
∂u1
These give the local expression of Y k along the first segment (u1 , 0):
∂Γk1j
1
Y k (u1 , 0) = Y k (0, 0) − Y j Γk1j u1 + Y l Γj1l Γk1j − Y j u21 + O(u31 ).
(0,0) 2 ∂u1 (0,0)
Next, we consider the second segment: transporting Y (u1 , 0) along the u2 -curve. By a
similar Taylor expansion, we get:
∂Y k 1 ∂2Y k
Y k (u1 , u2 ) = Y k (u1 , 0) + u2 + u22 + O(u32 ).
∂u2 (u1 ,0) 2 ∂u22 (u1 ,0)
As before, we next rewrite the derivative terms using Christoffel’s symbols. The parallel
transport equation along the u2 -curve gives:
∂Y k
= −Y j Γk2j
∂u2
∂2Y k l j k j
∂Γk2j
= Y Γ Γ
2l 2j − Y .
∂u22 ∂u2
∂Γk2j
k k j 1 l j k
Y (u1 , u2 ) = Y (u1 , 0) − Y Γk2j u2 + Y Γ2l Γ2j − Y j
u22 + O(u32 ).
(u1 ,0) 2 ∂u2 (u1 ,0)
Each term Y k (u1 , 0) above can then be expressed in terms of Y k (0, 0) and Γkij (0, 0)
using our previous calculations. It seems a bit tedious, but if we just keep terms up to
242 10. Curvatures of Riemannian Manifolds
∂Γk1j
k j k 1 l j k j
= Y (0, 0) − Y Γ1j u1 + Y Γ1l Γ1j − Y u2
(0,0) 2 ∂u1 (0,0) 1
∂Γk2j
j
− Y (0, 0) − Y (0, 0)Γ1l (0, 0)u1 Γk2j (0, 0) u2 − Y j (0, 0)
j l
u1 u2
∂u1 (0,0)
!
k
1 ∂Γ2j 3
+ Y l Γj2l Γk2j − Y j u22 + O |(u1 , u2 )| .
2 ∂u2 (0,0)
Next, we consider the parallel transport in another path, first along u2 , then along
u1 . Let Ye = Ye k ∂u∂ k be the local expression of the transported vector, then the local
expression of Ye k at (u1 , u2 ) can be obtained by switching 1 and 2 of that of Y k (u1 , u2 ):
∂Γk2j
1 el j k
Ye k (u1 , u2 ) = Ye k (0, 0) − Ye j Γk2j u2 + Y Γ2l Γ2j − Ye j
u2
(0,0) 2 ∂u2 (0,0) 2
∂Γk1j
j
e j e l
− Y (0, 0) − Y (0, 0)Γ2l (0, 0)u2 Γk1j (0, 0) u1 − Ye j (0, 0) u1 u2
∂u2 (0,0)
!
1 el j k ∂Γk1j 3
+ Y Γ1l Γ1j − Ye j u21 + O |(u1 , u2 )| .
2 ∂u1 (0,0)
Recall that Y (0, 0) = Ye (0, 0). By comparing Y (u1 , u2 ) and Ye (u1 , u2 ), we get:
Y k (u1 , u2 ) − Ye k (u1 , u2 )
∂Γk1l ∂Γk2l
3
= Y l Γj1l Γk2j − Γj2l Γk1j +
− u1 u2 + O |(u1 , u2 )|
∂u2 ∂u1 (0,0)
k 3
= R21l Yl u1 u2 + O |(u1 , u2 )| .
(0,0)
Therefore, the vector Rm(∂2 , ∂1 )Y at p measures the difference between the parallel
transports of those two different paths.
10.1.3. Symmetric properties and Bianchi identities. We will explore more about
the geometric meaning of the Rm(4,0) -tensor in the next section. Meanwhile, let’s discuss
some nice algebraic properties of this tensor. The Riemann curvature (4, 0)-tensor satisfies
some nice symmetric properties. The first two indices ij, and the last two indices kl
of the components Rijkl , and is symmetric if one swap the whole ij with the whole kl.
Precisely, we have:
10.1. Riemann Curvature Tensor 243
Proof. The fact that Rijkl = −Rjikl follows immediately from the definition of Rm(3,1) .
It only remains to show Rijkl = Rklij , then Rijkl = −Rijlk would follow immediately.
We prove it by picking geodesic normal coordinates at a fixed point p, so that
gij = δij , ∂k gij = 0, and Γkij = 0 at p.
Then at p, we have:
Rijkl − Rklij
p p
= gpl Rijk − gpj Rkli
∂Γpjk
!
∂Γpik
= gpl − + Γqjk Γpiq − Γqik Γpjq
∂ui ∂uj
p
∂Γpki
∂Γli q p q p
− gpj − + Γli Γkq − Γki Γlq
∂uk ∂ul
∂Γljk ∂Γlik ∂Γjli ∂Γjki
= − − + .
∂ui ∂uj ∂uk ∂ul
Note that in geodesic normal coordinates we only warrant Γkij = 0 at a point, which does
not imply its derivatives vanish at that point! Next we recall that
1 ∂gkp ∂gjp ∂gjk
Γljk = g lp + − ,
2 ∂uj ∂uk ∂up
and so at p we have:
∂Γljk 1 lp ∂ 2 gkp ∂ 2 gjp ∂ 2 gjk
2
∂ 2 gjl ∂ 2 gjk
1 ∂ gkl
= g + − = + − .
∂ui 2 ∂ui ∂uj ∂ui ∂uk ∂ui ∂up 2 ∂ui ∂uj ∂ui ∂uk ∂ui ∂ul
By permutating the indices, one can find out similar expressions for
∂Γlik ∂Γjli ∂Γjki
, , and .
∂uj ∂uk ∂ul
Then by cancellations one can verify that Rijkl − Rklij = 0 at p. Since Rijkl − Rklij is
tensorial, it holds true under any local coordinate system and at every point.
Another nice property of Rm(4,0) is the following pair of Bianchi identities, which
assert that the indices of the tensor exhibit some cyclic relations.
Proposition 10.3 (Bianchi Identities). The local components of Rm(4,0) satisfy the fol-
lowing: for any i, j, k, l, p, we have
R(ijk)l := Rijkl + Rjkil + Rkijl = 0.
∇(i Rjk)lp := ∇i Rjklp + ∇j Rkilp + ∇k Rijlp = 0.
Sketch of proof. Both can be proved using geodesic normal coordinates. Consider the
geodesic normal coordinates {ui } at a point p, then as in the proof of Proposition 10.2,
we have
∂Γljk ∂Γlik
Rijkl = − at p.
∂ui ∂uj
244 10. Curvatures of Riemannian Manifolds
By relabelling indices, we can write down the local expressions of Rjkil and Rkijl . By
summing them up, one can prove the first Bianchi identity by cancellations.
The second Bianchi identity can be proved in similar way. However, we have to be
careful not to apply Γkij (p) = 0 too early! Even though one has Γkij = 0 at a point p, it
does not warrant its derivatives equal 0 at p. The same for ∂k gij , which are all zero at p,
but it does not imply ∂l ∂k gij = 0 at p.
We sketch the proof and leave the detail for readers to fill in. Under geodesic normal
coordinates at p, we have
∇i Rjklp = ∂i Rjklp , at p.
Now consider
∂Γqjl
( !)
∂ ∂ ∂Γqkl m q m q
Rjklp = gqp − + Γkl Γjm − Γjl Γkm .
∂ui ∂ui ∂uj ∂uk
Use (7.15) again to write Γkij ’s in terms of derivatives of gij ’s. Be caution that to evaluate
∂
at p only after differentiation by ∂u i
. Get similar expressions for ∇j Rkilp and ∇k Rijlp at
p. One then should see all terms got cancelled when summing them up.
Exercise 10.2. Prove the second Bianchi identity using geodesic normal coordinates.
10.1.4. Isometric invariance. The Riemann curvature tensors (both types) can
be shown to be invariant under isometries, meaning that if Φ : (M, g) → (M f, ge) is an
∗ ∗g
isometry (i.e. Φ ge = g), then we have Φ Rm = Rm. To prove this, we first show that the
Levi-Civita connection is also isometric invariant.
∞
for any X, Y ∈ Γ (T M ). Here ∇ and ∇
e are the Levi-Civita connections for g and ge
respectively.
Idea of Proof. While it is possible to give a proof by direct computations using local
coordinates, there is a much smarter way of doing it. Here we outline the idea of proof
and leave the detail for readers to fill in. The desired result is equivalent to:
∇X Y = (Φ−1 )∗ ∇
e Φ X (Φ∗ Y ) .
∗
Then, one can verify that such D is a connection on M satisfying conditions (8.7) and
(8.8). This shows D must be the Levi-Civita connection of M by Proposition (8.14),
completing the proof.
10.1. Riemann Curvature Tensor 245
Now given that the Levi-Civita connection is isometric invariant, it follows easily that
Rm is too.
Φ∗ Rm
g = Rm
where both Riemann curvature tensors are of (4, 0)-type.
Example 10.6. As derivatives (hence curvatures) are local properties, the above results
also hold if Φ is just an isometry locally between two open subsets of two Riemannian
manifolds. The flat metric δ on Rn certainly has Rm = 0. Now for the torus Tn := Rn /Zn ,
it has an induced metric g from the covering map π : Rn → Tn (see Proposition 8.11)
which is a local isometry since π ∗ g = δ. This metric also has a zero Riemann curvature
tensor by Proposition 10.5. As such, we call this torus with such an induced metric the
flat torus.
Example 10.7. Consider the round sphere Sn , defined as x21 + · · · + x2n+1 = 1 in Rn+1 ,
with Riemannian metric given by the first fundamental form g = ι∗ δ. From standard Lie
theory, one can show the symmetry group SO(n + 1) acts transitively on Sn , meaning
that given any p, q ∈ Sn , there exists Φ : Rn+1 → Rn+1 such that Φ(p) = q.
R1221 Rm(∂1 , ∂2 , ∂2 , ∂1 )
K= 2 = 2 2 .
g11 g22 − g12 |∂1 | |∂2 | − g(∂1 , ∂2 )2
X1 = aX2 + bY2
Y2 = cX2 + dY2
Then, we have
Rm(X1 , Y1 , Y1 , X1 )
= Rm(aX2 + bY2 , cX2 + dY2 , cX2 + dY2 , aX2 + bY2 )
= ad Rm(X2 , Y2 , cX2 + dY2 , aX2 + bY2 )
+ bc Rm(Y2 , X2 , cX2 + dY2 , aX2 + bY2 )
= ad ad Rm(X2 , Y2 , Y2 , X2 ) + bc Rm(X2 , Y2 , X2 , Y2 )
+ bc ad Rm(Y2 , X2 , Y2 , X2 ) + bc Rm(Y2 , X2 , X2 , Y2 )
= (ad − bc)2 Rm(X2 , Y2 , Y2 , X2 ).
10.2. Sectional Curvature 247
2 2 2
− ac |X2 | + (ad + bc)g(X2 , Y2 ) + bd |Y2 |
2 2
= (ad − bc)2 |X2 | |Y2 | − g(X2 , Y2 )2 .
Remark 10.10. Therefore, one can also define the sectional curvature at p associated to
a plane (i.e. 2-dimensional subspace) Π in Tp M :
Kp (Π) := Kp (X, Y )
where {X, Y } is any basis for Π.
Remark 10.11. Note that Kp (X, Y ) is not a (2, 0)-tensor! It is evident from the above
result that K(2X, Y ) = K(X, Y ) 6= 2K(X, Y ).
Although the sectional curvature is essentially the Riemann curvature tensor re-
stricted on inputs of type (X, Y, Y, X), it is interesting that Rm(X, Y, Z, W ) itself can also
be expressed in terms of sectional curvatures:
Proposition 10.12. The Riemann curvature tensor Rm(4,0) is uniquely determined by its
sectional curvatures. Precisely, given any tangent vectors X, Y, Z, W ∈ Tp M , we have:
Rm(X, Y, Z, W )
= Rm(X + W, Y + Z, Y + Z, X + W ) − Rm(X + W, Y, Y, X + W )
− Rm(X + W, Z, Z, X + W ) − Rm(X, Y + Z, Y + Z, X)
− Rm(W, Y + Z, Y + Z, W ) + Rm(X, Z, Z, X) + Rm(W, Y, Y, W )
− Rm(Y + W, X + Z, X + Z, Y + W ) + Rm(Y + W, X, X, Y + W )
+ Rm(Y + W, Z, Z, Y + W ) + Rm(Y, X + Z, X + Z, Y )
+ Rm(W, X + Z, X + Z, W ) − Rm(Y, Z, Z, Y ) − Rm(W, X, X, W ).
Proof. Omitted. We leave it as an exercise for readers who need to wait for half an hour
for a morning minibus back to HKUST.
It is worthwhile to note that the result holds true when Rm is replaced by any
(4, 0)-tensor T satisfying all of the following:
• Tijkl = −Tjikl = −Tijlk = Tklij
• Tijkl + Tjkil + Tkijl = 0
Proof. Denote ∇ and Rm the Levi-Civita connection and the Riemann curvature tensor
of (M, g), and denote ∇ and Rm to be those of (Σ, g). Here we denote Σ := Σp (X, Y )
for simplicity.
The Gauss curvature of (Σ, g) is given by
Rm(X, Y, Y, X)
.
g(X, X)g(Y, Y ) − g(X, Y )2
As we have g = g when restricted on Tp Σ, it suffices to show:
Rm(X, Y, Y, X) = Rm(X, Y, Y, X).
We are only left to show the terms involving h vanish at p. Take an orthonormal frame1
{eα } of normal vectors to the surface Σ. Write h(∂u , ∂v ) = hα α
uv eα (and similarly for huu
α
and hvv ). Then, one can compute that:
g ∇u h(∂v , ∂v ) , ∂u = g ∇u hα = g hα
e
vv α , ∂ u vv ∇u eα , ∂u .
Here we have used the fact that eα ⊥ ∂u . Using this fact again, one can also show that
g ∇u eα , ∂u = −g(eα , ∇u ∂u ) = 0 at p.
Similarly, one can also show
g ∇v h(∂u , ∂v ) , ∂u = 0 at p.
As we have seen in the proof above, the key idea is to relate Rm and Rm. In the
above proof it suffices to consider inputs of type (X, Y, Y, X), yet it is not difficult to
generalize the calculation to relate Rm(X, Y, Z, W ) and Rm(X, Y, Z, W ) of any input
types. This in fact gives a generalized Gauss’s equation, which we leaves it as an exercise
for readers.
Exercise 10.4. Let (M, g) be a Riemannian manifold with Levi-Civita connection ∇,
and (Σ, g) be a submanifold of M with induced metric g and Levi-Civita connection
∇. Given any vector field X, Y ∈ T Σ, we denote
h(X, Y ) := (∇X Y )N = normal projection of ∇X Y onto the normal space N Σ.
Prove that for any X, Y, Z, W ∈ T Σ, we have
Rm(X, Y, Z, W )
= Rm(X, Y, Z, W ) + h(X, Z)h(Y, W ) − h(X, W )h(Y, Z).
1“Frame” typically means a set of locally defined vector fields which are linearly independent.
250 10. Curvatures of Riemannian Manifolds
Exercise 10.5. Prove the above symmetric properties and the first Bianchi identity
for h ∧ k.
Now if (M, g) has constant sectional curvature, then there exists a constant C such that
Kp (X, Y ) = C for any p ∈ M and any linearly independent vectors {X, Y } ⊂ Tp M . In
other words, we have
C
Rm(X, Y, Y, X) = C g(X, X)g(Y, Y ) − g(X, Y )2 = − (g ∧ g)(X, Y, Y, X).
2
Since Rm + C2 (g ∧ g) is a (4, 0)-tensor satisfying symmetric properties and the first
Bianchi’s identity, and it equals zero when acting on any (X, Y, Y, X), we can conclude
that Rm + C2 (g ∧ g) ≡ 0. To conclude, we have proved that if Kp (X, Y ) = C for any
p ∈ M and any linearly independent vectors {X, Y } ⊂ Tp M , then
Rm(X, Y, Z, W ) = C g(X, W )g(Y, Z) − g(X, Z)g(Y, W ) ,
Rijkl = C(gil gjk − gik gjl ).
Example 10.14. The most straight-forward example of metrics with constant sectional
curvature is the Euclidean space Rn with the flat metric δ. We have Rm ≡ 0. Consequently,
the n-torus Tn equipped with the quotient metric given by the covering π : Rn → Tn
also has 0 sectional curvature.
Example 10.15. The round sphere Sn of radius r, given by x21 + · · · + x2n+1 = r2 , has
constant sectional curvature r12 . The key reason is that geodesics are great circles of
radius r. Given any linearly independent vectors X, Y ∈ Tp Sn , the sectional surface
Σp (X, Y ) considered in Proposition 10.13 is formed by spreading out geodesics from p.
As each geodesic has the same curvature 1r , the principal curvatures of Σp (X, Y ) at p are
{ 1r , 1r }, showing that Σp (X, Y ) has Gauss curvature r12 at p.
This shows Rijkl = C(gil gjk − gik gjl ) for some constant C. Next we try to find out
what this C is. Consider the sectional curvature associated to {∂i , ∂j }, then we have:
1 Rijji
= K(∂i , ∂j ) = .
r2 gii gjj − gij gji
By the fact that Rijji = C(gii gjj − gij gji ) for constant sectional curvature metrics, we
must have C = r12 , and we conclude that:
1
Rijkl = (gil gjk − gik gjl ).
r2
Example 10.16. The hyperbolic space Hn under the upper-half space model:
Hn := {(x1 , · · · , xn ) ∈ Rn : x1 > 0}
10.2. Sectional Curvature 251
can be easily shown to have constant sectional curvatures by direct computations. Recall
that its Riemannian metric is given by:
Pn
dxi ⊗ dxi
g = i=1 2 .
x1
1
Under the global coordinates {xi }, the metric components gij = δ
x21 ij
forms a diagonal
matrix, so g ij
= x21 δij .
It follows that ∂k gij = −2x−3
1 δk1 δij .
1 1
Γkij = g kl ∂i gjl + ∂j gil − ∂l gij = x21 ∂i gjk + ∂j gik − ∂k gij
2 2
1
= x21 − 2x−3 −3 −3
1 δi1 δjk − 2x1 δj1 δik + 2x1 δk1 δij
2
1
= δk1 δij − δi1 δjk − δj1 δik .
x1
By the local expression:
∂Γljk ∂Γlik
Rijkq = gql − + Γpjk Γlip − Γpik Γljp ,
∂ui ∂uj
it is straight-forward to verify that Rijkl = C(g ∧ g)ijkl for some constant C > 0. This is
equivalent to saying that the metric has negative constant sectional curvature. We leave
it as an exercise for readers to complete the computations.
Exercise 10.6. Complete the computations of Rijkq for the hyperbolic space under
the upper-half plane model. Find out the C > 0 and the sectional curvature explicitly.
Also, consider the rescaled hyperbolic metric ge := αg, where α > 0 is a constant.
What is the sectional curvature of ge?
10.2.4. Curvature Operator. Another interesting notion about the Rm(4,0) -tensor
is the curvature operator. Recall that the tensor satisfies symmetric properties Rijkl =
−Rjikl = Rklij . If we group together i with j, and k with l, then one can regard Rm(4,0)
as a symmetric operator on ∧2 T M . Precisely, we define
252 10. Curvatures of Riemannian Manifolds
R : ∧2 T M → ∧2 T M
R(X ∧ Y ) := −Rm X, Y, g pk ∂p , g lq ∂q ∂k ∧ ∂l
and extend tensorially to all of ∧2 T M . The minus sign is to make sure that positive
curvature operator (to be defined later) would imply positive sectional curvature.
Furthermore, given the metric g, we can define an induced metric, still denoted by g,
on ⊗2 T M and ∧2 T M by the following way:
g(X ⊗ Y, Z ⊗ W ) := g(X, Z)g(Y, W ).
As X ∧ Y := X ⊗ Y − Y ⊗ X, one can easily verify that
g(X ∧ Y, Z ∧ W ) = (g ∧ g)(X, Y, Z, W ).
Then, one can also verify easily that R is a self-adjoint operator with respect to this
induced metric g on ∧2 T M , meaning that
g R(X ∧ Y ), Z ∧ W = g X ∧ Y, R(Z ∧ W ) .
Indeed, we have g R(X ∧ Y ), Z ∧ W = −Rm(X, Y, Z, W ). By standard linear algebra,
such an operator would be diagonalizable and that all its eigenvalues are real.
Exercise 10.7. Verify all the above claims, including
• R is well-defined;
• g(X ∧ Y, Z ∧ W ) = (g ∧ g)(X, Y, Z, W );
• g R(X ∧ Y ), Z ∧ W = Rm(X, Y, Z, W ); and
• R is self-adjoint with respect the inner product g on ∧2 T M .
We say (M, g) has positive curvature operator if all eigenvalues of R are positive at
every point in M . The curvature operator is related to sectional curvatures in a sense
that the later are the “diagonals” of R. As a matrix is positive-definite implies all its
diagonal entries are positive (NOT vice versa), so positive curvature operator implies
positive sectional curvatures. The converse is not true: CPn with the Fubini-Study metric
can be shown to have positive sectional curvatures (in fact K(X, Y ) is either 1 or 4), but
does not have positive curvature operator.
It had been a long-standing conjecture (recently solved in 2006) that what topology
a compact Riemannian manifold (M, g) must have if it has positive curvature operator. In
dimension 3, positive curvature operator implies positive Ricci curvature (to be defined
in the next section). In 1982, Richard Hamilton introduced the Ricci flow to show that
such a 3-manifold must be diffeomorphic to S3 with round metric or its quotient. The
key idea (modulo intensive technical detail) is to show that such a metric would evolve
under the heat equation to the one which constant sectional curvature (compared with
heat diffusion distributes temperature evenly in the long run), which warrants that such
a manifold must be diffeomorphic to S3 or its quotient.
Several years after the celebrated 3-manifold paper, Hamilton used the Ricci flow
again to show that if a compact 4-manifold (M, g) has positive curvature operator, then
it must be diffeomorphic to S4 or its quotient. He conjectured that the result holds for
any dimension. It was until 2006 that Christoph Boehm and Burkhard Wilking resolved
this conjecture completely, again using the Ricci flow (combined with some Lie algebraic
techniques).
10.3. Ricci and Scalar Curvatures 253
10.3.1. Ricci Curvature. We will define the Ricci curvature using both invariant
(global) notations and local coordinates. Let’s start with the invariant notations first
(which give clearer geometric meaning):
By the symmetry properties of Rm, it is clear that Ric is a symmetric tensor, i.e.
Ric(X, Y ) = Ric(Y, X).
Given a unit vector X ∈ Tp M , one can pick an orthonormal basis {ei } for Tp M such
that e1 = X, then we have:
n
X n
X
Ric(X, X) = Rm(ei , e1 , e1 , ei ) = K(ei , e1 ).
i=2 i=2
Therefore, the quantity Ric(X, X) is essentially the sum of sectional curvatures associated
to all orthogonal planes containing X.
We say that (M, g) has positive Ricci curvature at p ∈ M if Ric is positive-definite
at p, namely Ric(X, X) ≥ 0 for any X ∈ Tp M with equality holds if and only if X = 0.
If it holds true at every p ∈ M , we may simply say (M, g) has positive Ricci curvature.
Clearly, positive sectional curvature implies positive Ricci curvature. If one only has
Ric(X, X) ≥ 0 for any X ∈ Tp M without the equality conditions, we say (M, g) has
semi-positive (or non-negative) Ricci curvature at p.
Exercise 10.9. Suppose (M, g) has non-negative sectional curvature. Show that if
Ric ≡ 0, then Rm ≡ 0.
One nice way of proving it is to use geodesic normal coordinates. Fix a point p ∈ M ,
and pick an orthonormal basis {ei } ∈ Tp M , then there exists a local coordinate system
∂
{ui } such that ∂u i
= ei at p, and gij = 0 at p. By Exercise 10.8, the definition of Ric is
independent of such an orthonormal basis.
254 10. Curvatures of Riemannian Manifolds
Now we consider:
X
Rij := Ric(∂i , ∂j ) = Rm(ek , ∂i , ∂j , ek )
k
X X
= Rm(∂k , ∂i , ∂j , ∂k ) = Rkijk
k k
= δlk Rkijl = g lk Rkijl .
P
The last line was to convert k Rkijk into a tensorial quantity. Now that both Rij and
g lk Rkijl are tensorial (the latter is so because the summation indices k and l appear
as top-bottom pairs), so the identity Rij = g lk Rkijl holds under any local coordinates
covering p. Again since p is arbitrary, we have proved (10.1) holds globally on M .
It is possible to give a less “magical” (but more transparent) proof of (10.1). Write
∂
ek = Aik ∂u i
where {ek } is an orthonormal basis of Tp M and {ui } is any local coordinate
system. By orthogonality, we have
δij = g(ei , ej ) = Aki gkl Alj .
In matrix notations, we have I = [A][g][A]T , where Aki is the (i, k)-th entry of [A], then
we get [g] = [A]−1 ([A]T )−1 and [g]−1 = [A]T [A], i.e.
X
g ij = Aik Ajk .
k
Now consider
X
Rij = Rm(ek , ∂i , ∂j , ek )
k
X
= Rm(Apk ∂p , ∂i , ∂j , Aqk ∂q )
k,p,q
X X
= Apk Aqk Rpijq = g pq Rpijq
k,p,q p,q
as desired.
10.3.2. Scalar Cuvature. The scalar curvature R is the trace of the Ricci tensor, in
a sense that
Xn
R := Ric(ei , ei )
i=1
where {ei } is an orthonormal frame of the tangent space. It can be shown that R
is independent of the choice of the orthonormal frame, and has the following local
expression
(10.2) R = g ij Rij
Exercise 10.10. Show that the definition of R does not depend on the choice of the
orthonormal frame {ei }ni=1 , and prove (10.2).
Using the second Bianchi identity, one can prove the following identity relating the
Ricci curvature and scalar curvature.
One immediate consequence of the contracted Bianchi identity is that the 2-tensor
Ric − R2 g is divergence-free:
R 1 1 1
∇i Rij − gij = ∇i Rij − gij ∇i R = ∂i R − ∂i R = 0.
2 2 2 2
R
The 2-tensor G := Ric − 2g is called the Einstein tensor, named after Albert Einstein’s
famous equation:
R
Rµν − gµν = Tµν
2
where the tensor Tµν described stress and energy. Here we use µ and ν for indices
instead of i and j as physicists use the former. We will learn in the next subsection
R p that a
metric satisfying the Einstein’s equation is a critical metric of the functional M R det[g],
known as the Einstein-Hilbert’s functional.
Inspired by the (vacuum) Einstein’s equation (where T = 0), mathematicians called
a Riemannian metric g satisfying the equation below an Einstein metric:
Rij = cgij , where c is a real constant.
A metric g with constant sectional curvature must be an Einstein metric. To see this, let’s
suppose
Rijkl = C(gil gjk − gik gjl ).
Then by tracing both sides, we get:
Rjk = g il Rijkl = C(g il gil gjk − g il gik gjl ) = C(ngjk − δlk gjl ) = C(n − 1)gjk .
When dim M ≥ 3, a connected Riemannian metric satisfying Rij = f gij , where f is
a smooth scalar function, must be an Einstein metric. To show this, we first observe that
f=R n:
R = g ij Rij = g ij f gij = nf.
Then, the contracted Bianchi identity shows
1 R 1 1
∂j R = ∇i Rij = ∇i gij = gij ∇i R = ∂j R.
2 n n n
When n > 2, it implies ∂j R = 0 for any j, and hence R is a constant.
The proof obviously fails in dimension 2. In contrast, it is always true that Rij = R2 gij
when n = 2, for any Riemannian metric g. It can be seen easily by combinatoric
inspections:
256 10. Curvatures of Riemannian Manifolds
An Einstein metric must have constant scalar curvature. The proof is easy: if
Rij = cgij , then R = g ij Rij = nc. It was a fundamental question of what manifolds
admit a metric with constant scalar curvature. In dimension 2, it was the Uniformization
Theorem. Any simply connected Riemann surface (i.e. complex manifold with real
dimension 2) must be biholomorphic (i.e. conformal) to one of the standard models with
constant curvature: disc, plane, sphere (corresponding to negative, zero, and positive
curvatures respectively).
In higher dimensions (i.e. dim M ≥ 3), the positive curvature case was known as the
Yamabe problem, named after Hidehiko Yamabe. The problem was progressively resolved
by Neil Trudinger, Thierry Aubin, and finally by Richard Schoen in 1984, concluding that
on any compact manifold (M, g) of dimension n ≥ 3 with positive scalar curvature, there
exists a smooth function f on M such that the conformally rescaled metric ge = ef g has
constant positive scalar curvature.
k ∂
the given Riemannian metric g. For instance, consider S = Sij dui ⊗ duj ⊗ ∂uk , and
T = Tijk dui ⊗ duj ⊗ ∂u∂ k , then we define:
g(S, T ) := g ip g jq gkr Sij
k r
Tpq .
It is a globally defined scalar function on M , and such a definition is independent of local
k k
P
coordinates. In the Euclidean case where gij = δij , then δ(S, T ) = i,j,k Sij Tij which
appears like the dot product.
We can then make sense of two tensor fields (of the same type) being orthogonal to
each other. For example, one can check:
R R R R
g Ric − g, g = g ik g jl Rij − gij gkl = g ij Rij − g ij gij = R − · n = 0,
n n n n
R
so Ric − ng and g are orthogonal.
Exercise 10.12. Show that g ∧ g and (Ric − R n g) ∧ g are orthogonal to each other
(with respect to the inner produced inherited from g).
Proof. Each two-form is spanned by dui ∧ duj with i < j. Hence we have
n(n − 1)
dim ∧2 T ∗ M = C2n = .
2
Similarly, given a vector space V of dimension N , we have
N (N + 1)
dim(V ⊗S V ) = C2N + N = .
2
Hence, we have
n(n−1) n(n−1)
Cn 2 2 +1
dim E = C2 2 = ,
2
as desired.
The dimension of B can be found in a trickier way. We consider a map known Bianchi
symmetrization:
b : E → ⊗4 T ∗ M
1
b(T )ijkl := (Tijkl + Tjkil + Tkijl ).
3
Then, B = ker(b), hence dim B = dim E − dim Im (b) by elementary linear algebra.
One can show that the image Im (b) is in fact ∧4 T ∗ M . To show b(T ) ∈ ∧4 T ∗ M , we
observe that Tikjl + Tkjil = −Tjkil − Tkijl , so that
3b(T )jikl = Tjikl + Tikjl + Tkjil = −Tijkl − Tjkil − Tkijl = −3b(T )ijkl ,
n(n − 1)(n2 − n + 2) 1 2 2
dim B = dim E − dim ∧4 T ∗ M = − C4n = n (n − 1).
8 12
L : S 2 (T ∗ M ) → B
h 7→ h ∧ g
Hence, we must have hij = 0 for any i, j. As h is tensorial, this holds true under any
local coordinate system.
In fact, the map L is adjoint to the trace map. Define a map:
Tr : B → S 2 (T ∗ M )
Tijkl (dui ∧ duj ) ⊗S (duk ⊗ dul ) 7→ g il Tijkl duj ⊗S duk
Then one can prove the following useful identity (whose proof is left as an exercise):
Exercise 10.13. Given any symmetric 2-tensor h, and a 4-tensor T ∈ B, we have:
g(T, L(h)) = g(4 Tr(T ), h).
As a result, the Weyl tensor is orthogonal to the image of L.
Now the key observation is that L is in fact bijective when dim M = 3. It is because
S 2 (T ∗ M ) and B both have the same dimension (equal to 6). Since W ∈ B, there exists a
symmetric 2-tensor h such that L(h) = W . Then, the fact that W = 0 follows easily from
Exercise 10.13:
g(W, W ) = g(W, L(h)) = 4g(Tr(W ), h) = 0
by the fact that Tr(W ) = 0. This clearly shows W = 0.
Consequently, in dimension 3, the Riemann curvature tensor can be expressed as:
R R
Rm = g ∧ g + Ric − g ∧ g.
12 3
An immediate corollary is that in dimension 3, the Ricci tensor Ric determines Rm. Also,
any Einstein metric (in dimension 3 only!) must have constant sectional curvature.
In dimension 4 or above, the Weyl tensor may not be 0. But being trace-free, it is
orthogonal to all 4-tensors of type h ∧ g (where h is symmetric 2-tensor). In particular, it
is orthogonal to both
R 1 R
(g ∧ g) and Ric − g ∧ g,
2n(n − 1) n−2 n
making the following an orthogonal decomposition:
R 1 R
Rm = g ∧ g+ Ric − g ∧ g + W.
2n(n − 1) n−2 n
The Weyl tensor can also be a conformal invariant: consider ge = ef g where f is a smooth
scalar function. It can be shown that the Weyl tensors are related by W (e g ) = ef W (g).
The proof is somewhat computational (either using geodesic normal coordinates or
Cartan’s notations). One needs to find out the how the Rm, Ric and R of ge are related to
those of g.
260 10. Curvatures of Riemannian Manifolds
10.4.1. Variation Formula for Ricci Tensor. Let’s first start with the Ricci tensor.
∂ ∂
Given that ∂t gij (t) = vij (t). We want to find a nice formula for ∂t Rij where Rij :=
Ricg(t) (∂i , ∂j ). Again we use geodesic normal coordinates at a fixed point p. Recall that
k
Rij = Rkij = ∂k Γkij − ∂i Γkkj + quadratic terms of Γkij ’s.
Although Γkij (p) = 0 does not warrant its derivative is 0 at p, we still have
∂
(quadratic terms of Γkij ’s) = 0
∂t
at p by the product rule. Hence, we may focus on the the first two terms when computing
∂
∂t Rij . We consider:
∂ ∂ ∂ k
∂k Γkij = Γ
∂t ∂uk ∂t ij
∂ 1 kl
= g (∇i vjl + ∇j vil − ∇l vij )
∂uk 2
1 ∂
= g kl (∇i vjl + ∇j vil − ∇l vij ) .
2 ∂uk
Here we have use the fact that ∂k g kl = 0 at p for geodesic normal coordinates. Next note
that
∇k ∇i vjl = ∂k (∇i vjl ) + terms with Christoffel symbols,
so under geodesic normal coordinates we have
∇k ∇i vjl = ∂k (∇i vjl ).
10.4. Variation Formulae of Curvatures 261
In particular, if the metric g(t) deforms along the direction of −2Ric(g(t)), i.e.
∂
gij = −2Rij ,
∂t
then the scalar curvature evolves by
∂ 2
R = −∆(−2R) − g(−2Ric, Ric) − 2∇i ∇j Rij = ∆R + 2 |Ric|
∂t
262 10. Curvatures of Riemannian Manifolds
where we have used the contracted Bianchi identity ∇j Rij = 12 ∂i R. Using parabolic
maximum principle, one can then show on a compact manifold if R ≥ C at t = 0, then
∂
it remains so for all t > 0 under the evolution ∂t gij = −2Rij . This deformation of the
metric g(t) is called the Ricci flow, which is the major tool for resolving the Poincaré
conjecture.
10.4.3. Variation Formula for Determinants. Next we want to compute the varia-
tion formula for det[gij ]. The following general result is very useful:
Lemma 10.20. Let A(t) be a time-dependent invertible n × n matrix with entries denoted
by Aij , then we have:
∂ −1 ∂A
(10.4) det A(t) = Tr A det A(t),
∂t ∂t
where Aij is the (i, j)-th entry of A−1 .
Proof. It is best be done using differential forms. Let {ei }ni=1 be the standard basis of
Rn . Then we have:
n
X ∂A ∂
Ae1 ∧ · · · ∧ Aei−1 ∧ ei ∧ Aei+1 ∧ · · · ∧ Aen = det(A) e1 ∧ · · · ∧ en .
i=1
∂t ∂t
Next we prove the following general result then the proof is completed. For any invertible
n × n matrix A, and any n × n matrix B, we have:
n
X
Ae1 ∧ · · · ∧ Aei−1 ∧ Bei ∧ Aei+1 ∧ · · · ∧ Aen = Tr(A−1 B) det(A) e1 ∧ · · · ∧ en .
i=1
To show this, we first prove the above holds for diagonalizable matrices A, which is
dense in the set of invertible matrices. Let P = [Pij ] be an invertible matrix such that
A = P DP −1 where D = diag(λ1 , · · · , λn ). Then, we have
n
X
Ae1 ∧ · · · ∧ Aei−1 ∧ Bei ∧ Aei+1 ∧ · · · ∧ Aen
i=1
n
X
= P DP −1 e1 ∧ · · · ∧ P (P −1 BP )P −1 ei ∧ · · · ∧ P DP −1 en
i=1
n
X
= det(P ) DP −1 e1 ∧ · · · ∧ (P −1 BP )P −1 ei ∧ · · · ∧ DP −1 en
i=1
n
X
= det(P ) det(D) P −1 e1 ∧ · · · ∧ D−1 (P −1 BP )P −1 ei ∧ · · · ∧ P −1 en
i=1
10.4. Variation Formulae of Curvatures 263
∂
As a corollary, given that g(t) satisfies ∂t gij = vij , then we have
∂
det[gij ] = Tr([g]−1 [v]) det[g] = g ij vij det[g] = (Trg v) det[g].
∂t
One immediate consequence of the above variation formula is that the mean cur-
vature H of a Euclidean hypersurface Σ (with boundary C) must be 0 if it minimizes
the area among all smooth hypersurface with the boundary C. To see this, we let
Σ0 be such a hypersurface, and Σt is a 1-parameter family of hypersurfaces such that
Area(Σ0 ) ≤ Area(Σt ) for any t.
Denote Ft (u1 , · · · , un ) to be the local parametrization of Σt , and gij (t) be the first
fundamental form. Suppose ∂F ∂t = f ν where ν := νt is the unit normal for Σt , then one
can check that
∂gij ∂ ∂F ∂F ∂ν ∂F ∂F ∂ν
= , = f , + ,f .
∂t ∂t ∂ui ∂uj ∂ui ∂uj ∂ui ∂uj
∂
Here we have used the fact that ν and ∂ui are orthogonal. Recall that
∂ν ∂F
= −hki ,
∂ui ∂uk
we conclude that
∂gij
= −2f hij ,
∂t
and hence
∂p 1 p
det[g] = p (−2f g ij hij ) det[g] = −f H det[g].
∂t 2 det[g]
264 10. Curvatures of Riemannian Manifolds
10.4.4. Deriving the vacuum Einstein equation. Finally (after a short detour to
minimal surfaces), we get back on p
deriving the Einstein equation. The essential task is to
∂
derive the variation formula for R det[g]. Suppose ∂t gij = vij , then
∂ p p 1 p
R det[g] = −∆Trg v − g(v, Ric) + ∇i ∇j vij
det[g] + R · Trg v · det[g]
∂t 2
R p
= −∆Trg v − g(v, Ric) + ∇i ∇j vij + g(v, g) det[g]
2
R p
= −∆Trg v − g v, Ric − g + ∇i ∇j vij det[g].
2
Consequently, we have
Z
d R p
LEH (g(t)) = −∆Trg v − g v, Ric − g + ∇i ∇j vij det[g] du1 ∧ · · · ∧ dun .
dt M 2
Assuming M has no boundary, any divergence terms such as ∇i Tij or ∇i T ij , etc.
must have integral 0 by the Stokes’ Theorem. This is can be shown by the following nice
observation:
Exercise 10.14. Suppose X is a vector field on a Riemannian manifold (M, g)
without boundary. Denote the volume element by
q
dµg := det[gij ] du1 ∧ · · · ∧ dun .
Show that
d(iX dµg ) = ∇i X i dµg .
Hence, by Stokes’ Theorem, we get
Z Z
i
∇i X dµg = d(something) = 0.
M M
Therefore, the variation formula for the LEH -functional can be simplified as:
Z
d R
LEH (g(t)) = − g v, Ric − g dµg .
dt M 2
If we need g(0) to minimize LEH under any variations v, then it is necessary that
R
Ric − g ≡ 0
2
which is exactly the vacuum Einstein’s equation.
By taking the trace with respect to g on both sides, we get:
R
R − n = 0.
2
If n > 2, we then have R = 0, and so Ric ≡ 0. Therefore, the solution to the vacuum
Einstein’s equation is necessarily Ricci-flat (but may not be Riemann-flat).
Exercise 10.15. Consider the following modified Einstein-Hilbert’s functional:
Z
LeEH (g) := R − 2Λ + ϕ(g) dµg
M
where Λ is a real constant, and ϕ(g) is a scalar function depending on the metric g.
Show that is g minimizes this functional among all variations of g, then it is
necessary that
R
Ric − g + Λg = T
2
where T is a symmetric 2-tensor depending on ϕ and its derivatives.
Chapter 11
「隻手換乾坤,拓撲知曲率。」
丘成桐
A typical kind of questions that geometers and topologists would like to address is
that given some curvature conditions (note that curvatures are local properties), what
can one say about the global properties of the manifold (such as diameters, compactness,
topological type, etc.)?
There are two fundamental analytical tools for addressing these kind of questions,
namely the existence of Jacobi fields and the second variations of arc-lengths. The
Riemann curvature tensors play an important role in these two tools. They are used to
transfer local information about curvatures to some more global information (such as
diameter of a manifold).
One neat result is the following theorem due to Bonnet and Myers: which says that
if (M n , g) is a complete Riemannian manifold with Ric ≥ k(n − 1)g on M where k > 0 is
a positive constant, then one has diam(M, g(t)) ≤ √πk , and consequently M is compact
and has finite fundamental group π1 (M ). We will prove the theorem in Section 11.2.
The equality case of Bonnet-Myers’ Theorem was addressed by S.-Y. Cheng (Professor
Emeritus of HKUST Math) in 1975, who proved the equality holds if and only if (M n , g)
is a isometric to the round sphere with constant sectional curvature k.
In laymen terms, a Jacobi field J is a vector field whose integral curves are all
geodesics. If there are infinitely many geodesics connecting p and q, then there would
exist a Jacobi field J such that J(p) = J(q) = 0. An intuitive example is the round
2-sphere with p and q being the north and south poles. There is a family of great
∂ ∂
semi-circles connecting them and the Jacobi field is ∂ϕ (or ∂θ in PHYS convention).
Another fundamental result relating Jacobi fields and geodesics is that the existence
question of non-trivial Jacobi field J with J(p) = J(q) = 0 is (roughly) equivalent to
whether (expp )∗ is singular at the vector tX corresponding to q. In this case, if one
connects p and q by a unit-speed geodesic γ(t) (say γ(0) = p and γ(L) = q), then one can
use the above-mentioned equivalence to show that {γ(t)}t>0 is no longer a minimizing
geodesic when t > L. This can give an upper bound L on the distance d(p, q), and by
repeating the argument on arbitrary p and q, we can estimate the diameter (maximal
distance) of the manifold (M, g).
267
268 11. Curvatures and Topology
Rm(V, T )T = ∇V ∇T T − ∇T ∇V T − ∇[V,T ] T.
∂
We have ∇V ∇T T = ∇ ∂γ ∇T T = 0 since ∇T T = 0 holds for any s. Also, by V = γ∗ ∂s
∂s
∂
and T = γ∗ ∂t , we can see that [V, T ] = 0. The only survival term is the second one, and
further observe that ∇V T − ∇T V = [V, T ] = 0, so we get
Rm(V, T )T = −∇T ∇T V.
Note that (11.1) is a second-order ODE. Once we prescribed V (0) and ∇γ̇ V at t = 0,
then there exists a unique Jacobi field along γ with these initial data. The solution
space is a vector space by the linearity of the equation (11.1). The solution space is
parametrized by V (0) and ∇γ̇ V t=0 , hence is 2n-dimensional.
Example 11.2. Two easy examples of Jacobi fields are V = γ̇, and W = tγ̇. The first
one is obvious by the geodesic equation ∇γ̇ γ̇ = 0, and Rm(γ̇, γ̇)γ̇ = 0 by the alternating
property of Rm. For W = tγ̇, we also have Rm(W, γ̇)γ̇ = 0 by the same reason. We can
then show
∇γ̇ ∇γ̇ (tγ̇) = ∇γ̇ (γ̇ + t∇γ̇ γ̇ ) = ∇γ̇ γ̇ = 0.
| {z }
=0
The Jacobi fields γ̇ and tγ̇ are merely tangent vector fields so that their integral
curves are just reparametrization of the geodesic. The shape of the geodesic along these
variations is unchanged, and so they are not interesting examples. However, they are
essentially the “only” tangential Jacobi fields, since we can show any Jacobi field V along
γ can be decomposed into:
V = V ⊥ + aγ̇ + btγ̇
for some constants a and b, V ⊥ ⊥ γ̇, and still we have V ⊥ being a Jacobi field. The
argument is as follows:
11.1. Jacobi Fields 269
Consider the inner product g V (t), γ̇(t) , we will show that it is a linear polynomial
of t using the fact that ∇γ̇ γ̇ = 0:
d2
2
g V (t), γ̇(t)
dt
= g ∇γ̇ ∇γ̇ V, γ̇
= −g Rm(V, γ̇)γ̇, γ̇)
= −Rm(V, γ̇, γ̇, γ̇) = 0.
This shows g(V (t), γ̇(t)) = a + bt for some constants a and b. Then, it is easy to show
g V − aγ̇ − btγ̇, γ̇) = 0
and hence V − aγ̇ − btγ̇ ⊥ γ̇. Here we assume for simplicity that γ is arc-length
parametrized. By linearity of (11.1), V ⊥ is still a Jacobi field along γ, and it is normal to
the curve γ.
Consequently, the space of Jacobi fields has an orthogonal decomposition into
tangential and normal subspaces. The tangential subspace has dimension 2 spanned by γ̇
and tγ̇ (note that they are linearly independent as functions of t, even thought they are
parallel vectors pointwise), and the normal subspace has dimension 2n − 2.
Exercise 11.1. Suppose a Jacobi field V along a geodesic γ is normal to the curve
at two distinct points. What can you say about V ?
11.1.2. Jacobi Fields on Spaces with Constant Curvatures. Since the Jacobi
field equation (11.1) involves the curvature term, it is expected that one can find some
explicit solutions if the Riemann curvature term is nice. Consider a complete Riemannian
manifold (M, g) with constant sectional curvature C, i.e.
2 2
Rm(X, Y, Y, X) = C |X| |Y | − g(X, Y )2
for any X, Y ∈ Tp M .
Now given an arc-parametrized geodesic γ(t) : [0, T ] → M , and fix a unit vector
E0 ∈ Tγ(0) M such that E0 ⊥ γ̇(0), and extend this vector by parallel transport along γ,
i.e.
∇γ̇ E(t) = 0 and E(0) = E0 .
We want to find a Jacobi field along γ that is the form of V (t) = u(t)E(t) where u(t) is
a scalar function. It turns out that u(t) will satisfy a hand-solvable ODE if (M, g) has
constant sectional curvature.
Consider the Jacobi equation:
0 = ∇γ̇ ∇γ̇ (uE) + Rm(uE, γ̇)γ̇
= ∇γ̇ u̇E + uRm(E, γ̇)γ̇
= üE + uRm(E, γ̇)γ̇.
Here we have used the fact that ∇γ̇ E = 0. Taking inner product with E on both sides,
we get
0 = ü + ug Rm(E, γ̇)γ̇, E .
2
Here we have used the fact that |E(t)| = 1 by the property of parallel transport. Note
also that g(E(t), γ̇(t)) = 0 since it is so at t = 0, we have Rm(E, γ̇, γ̇, E) = C and we get
that
ü + Cu = 0.
270 11. Curvatures and Topology
Note that 2 2
2 d 2 V
|∇γ̇ V | − |V | = |∇γ̇ V | − g ∇γ̇ V, ≥0
dt |V |
by Cauchy-Schwarz’s inequality. We can conclude that as long as |V | > 0, we have
d2
|V | ≥ −C |V | .
dt2
Therefore, we can then derive comparison inequality between Jacobi fields of a
constant sectional curvature metric and a general metric with sectional curvature bounded
from above.
The key idea is that for any non-negative function f (t) that satisfies the inequalities
f 00 (t) + Cf (t) ≥ 0 for t > 0 , f (0) = 0, and f 0 (0) > 0,
then one can claim that
f 0 (0) √ π
√C sin(t C) for any t ∈ [0, √C ]
if C > 0
f (t) ≥ f 0 (0)t for any t ≥ 0 if C = 0 .
f 0 (0) √
sinh(t −C) for any t ≥ 0 if C < 0
√
−C
11.1. Jacobi Fields 271
To prove this, let u(t) be the solution to ü + Cu = 0 with u(0) = 0 and u0 (0) = f 0 (0), i.e.
f 0 (0) √
√C sin(t C)
if C > 0
u(t) = f 0 (0)t if C = 0 .
f 0 (0) √
sinh(t −C) if C < 0
√
−C
Sturm-Liouville’s comparison theorem of ODEs then asserts that f (t) ≥ u(t) as long as
u(t) > 0. Now let f (t) = |V (t)| where V (t) is a Jacobi field along γ such that V (0) = 0,
if one can argue that f 0 (0) > 0, then we conclude that
f 0 (0) √ π
√C sin(t C) for any t ∈ [0, √C ]
if C > 0
(11.2) 0
|V (t)| ≥ f (0)t for any t ≥ 0 if C = 0 .
f 0 (0) √
sinh(t −C) for any t ≥ 0 if C < 0
√
−C
As a corollary, if the sectional curvature is non-positive, any non-trivial Jacobi field V (t)
along γ(t) with V (0) = 0 would not vanish again for any t > 0.
We are left to show f 0 (0) > 0. Since V (0) = 0, we cannot compute f 0 (0) directly.
Consider the limit quotient:
|V (t)| − |V (0)| V (t)
f 0 (0) = lim+ = lim+ .
t→0 t t→0 t
We leave it as an exercise
for readers to show that for t > 0 sufficiently small, we in fact
have V (t) = t ∇γ̇ V γ(0) , then it completes the proof.
11.1.4. Conjugate Points and Jacobi Fields. One fundamental theorem about
Jacobi fields and the exponential maps is that the existence of Jacobi field V (t) along a
curve γ(t) : [0, T ] → M with V (0) = V (T ) = 0, is equivalent to (expγ(0) )∗ being singular
at the point corresponding to γ(T ). In this subsection, let’s make this relation more
precise. We first introduce:
∂
Example 11.4. The vector field (sin ϕ) ∂θ (in MATH spherical coordinates) on the stan-
dard 2-sphere is a Jacobi field vanishing at the north and south poles. Therefore, the
north and south poles are conjugate point to each other along any great circle passing
through them.
272 11. Curvatures and Topology
Proof. Before giving the proof, we first observe the following key fact. Let γs (t) be a
family of geodesic defined by:
γs (t) := expp t(T + sW )
where W is any vector in Tp M .
Then, γ0 = γ. The Jacobi field generated by this family is given by:
∂γs (t) ∂
(11.3) V (t) := = expp tT + s(tW )) = (expp )∗tT (tW ).
∂s s=0 ∂s s=0
The last step follows from the fact that the derivative is the directional derivative of expp
at tT along tW . In particular, V (q) = (expp )∗t0 T (t0 W ).
To show the (=⇒)-part, we consider a basis {Ei }ni=1 of Tp M , and consider the
family of geodesics γsi (t) := expp t(T + sEi ) which generates the Jacobi fields Vi (t) =
(expp )∗tT (tEi ) according to the above computation. These Jacobi fields are linearly
independent because (expp )∗tT is invertible for small t ≥ 0: suppose there are constants
λi ’s such that
Xn
λi Vi (t) ≡ 0,
i=1
then we have
Xn
(expp )∗tT t λi Ei ≡ 0.
i=1
Pick a small τ > 0 such that (expp )∗τ T is invertible, we have
n
X n
X
τ λi Ei = 0 =⇒ λi Ei = 0.
i=1 i=1
By linear independence of Ei ’s, we get λi = 0 for any i. Therefore, {Vi (t)}ni=1 is a basis
of Jacobi fields that vanish at p (which is an n-dimensional vector space).
To prove the desired claim, we suppose otherwise that there exists a non-zero Jacobi
field V along γ such that V (p) = V (q) = 0 but (expp )∗t0 T is invertible. From above, V (t)
must be spanned by {Vi (t)}ni=1 :
n
X
V (t) ≡ ci Vi (t).
i=1
and consequently
n
X
t0 ci Ei = 0 =⇒ ci = 0 for any i.
i=1
It is a contradiction to the fact that V (t) 6≡ 0.
11.1. Jacobi Fields 273
The (⇐=)-part is easier: suppose (expp )∗t0 T is singular, and in particular, there exists
W 6= 0 in Tp M such that
(expp )∗t0 T (W ) = 0.
Then, the Jacobi field V (t) defined in (11.3) satisfies:
V (q) = (expp )∗t0 T = 0.
By invertibility of (expp )∗tT for t > 0 small, V (t) is non-zero Jacobi field. It completes
the proof.
274 11. Curvatures and Topology
T
is that ∇T |T | = 0, and if we assume γ0 has constant speed, then ∇T T = 0 along γ.
However, to determine whether it is a local minimum, we need to consider the second
derivative. In this section, we will derive the second variation formula of arc-length and
discuss its applications.
where we have used the fact that [S, T ] = 0. By Pythagoreas’ Theorem, we have
2 2
|∇T S| − g(∇T S, T )2 = (∇T S)N ,
and as ∇T T = 0, we can also show
∇T S N = ∇T (S − g(S, T )T ) = ∇T S − g(∇T S, T )T = (∇T S)N .
11.2. Index Forms 275
Remark 11.7. The Rm(S, T, T, S) term in (11.4) also equals Rm(S N , T, T, S N ) since
Rm(T, T, T, T ) = 0.
In particular, for any V ∈ Vγ , the second variation of L(γ) along the variation field V
is given by I(V, V ). When γ is minimizing geodesic, it is necessary that I(V, V ) ≥ 0 for
any V ∈ Vγ . In other words, if we want to show a certain geodesic γ is not minimizing,
one needs to construct a vector field V ∈ Vγ such that I(V, V ) < 0.
When V ∈ Vγ is a C ∞ Jacobi field and W ∈ Vγ is any C ∞ vector field, one can use
integration by parts and show:
Z b
t=b−
I(V, W ) = g(∇T V, W ) t=a+ − g ∇T ∇T V + Rm(V, T )T, W dt = 0.
a
where we have used the fact that W (a) = W (b) = 0 and ∇T ∇T V + Rm(V, T )T = 0.
Note that if V and W are merely piecewise smooth on [a, b], say they are smooth on
[a, c) and (c, b], then we have
t=c− t=b−
I(V, W ) = g(∇T V, W ) t=a+ + g(∇T V, W ) t=c+ .
11.2.2. Geodesics Beyond Conjugate Points. One fundamental fact about geodesics
is that it is never minimizing if γ has an interior point conjugate to the starting point.
Proof. By Proposition 11.5, the given condition implies there exists a non-zero Jacobi
field V (t) defined on [a, c] such that V (a) = V (c) = 0. We can assume that V is normal
to γ̇.
Then, we extend V to the whole curve γ : [a, b] → M as follows:
(
V (t) if t ∈ [a, c]
V (t) :=
e .
0 if t ∈ (c, b]
276 11. Curvatures and Topology
Note that Ve (t) is not smooth. We are going to search for a C ∞ vector field W defined on
[a, b] and a small ε > 0 such that
I(Ve + εW, Ve + εW ) < 0.
Consider that
I(Ve + εW, Ve + εW ) = I(Ve , Ve ) + 2εI(Ve , W ) + ε2 I(W, W ).
Let’s compute each term one-by-one:
t=c− t=b−
I(Ve , Ve ) = g(∇T Ve , Ve ) t=a+ + g(∇T Ve , Ve ) t=c+
= g(∇T V (c), V (c)) − g(∇T V (a), V (a)) + g(∇T 0, 0) − g(∇T 0, 0)
=0
t=c− t=b−
I(Ve , W ) = g(∇T Ve , W ) t=a+ + g(∇T Ve , W ) t=c+
= g(∇T V (c), W (c)) − g(∇T W (a), W (a))
In order for W ∈ Vγ , we need W (a) = W (b) = 0. If we can construct W ∈ Vγ such that
W (c) = −∇T V (c), then we will have I(Ve , W ) < 0. Regardless of the sign of I(W, W ),
we will have
I(Ve + εW, Ve + εW ) = 2εI(Ve , W ) + ε2 I(W, W ) < 0
for sufficiently small ε > 0.
One can use a bump function to construct such a vector field W . Let ρ : [a, b] →
[0, 1] be a C ∞ function such that ρ(a) = ρ(b) = 0 and ρ(c) = 1. Take any paral-
lel unit vector field E(t) which is normal to γ̇(t). This vector field exists by parallel
transporting −∇T V (c) along the curve γ. Let X(t) be this parallel transport, then
W (t) := ρ(t)X(t)N is a smooth vector field in Vγ such that W (c) = −∇T V (c)N and so
2
I(Ve , W ) = − ∇T V (c)N < 0, completing the proof.
π
Proof. We claim that no minimizing geodesic has length greater than √
k
. Suppose
otherwise and γ(t) : [0, L] → M is a unit-speed geodesic with L > √πk .
Construct a parallel orthonormal frame {Ei (t)}ni=1 such that E1 (t) = γ̇(t), then for
each i we define π
Vi (t) := sin t Ei (t),
L
and consider its index form I(Vi , Vi ). Since each Ei (t) is parallel, we have
π 2
π π π
∇γ̇ Vi = cos t Ei (t) and ∇γ̇ ∇γ̇ Vi = − sin t Ei (t).
L L L2 L
11.2. Index Forms 277
Hence, we have
Z L
I(Vi , Vi ) = − g(∇γ̇ ∇γ̇ Vi , Vi ) + Rm(Vi , γ̇, γ̇, Vi ) dt
0
Z L 2
π 2 π
2 π
= 2
sin t − sin t Rm(Ei , E1 , E1 , Ei ) dt.
0 L L L
Summing up i over 2 to n, we get
n Z L n
!
(n − 1)π 2 X
2 π
X
I(Vi , Vi ) = sin t − Rm(Ei , E1 , E1 , Ei ) dt.
i=2 0 L L2 i=2
Note that by the given condition about the Ricci curvature, we have
X n
Rm(Ei , E1 , E1 , Ei ) = Ric(E1 , E1 ) ≥ (n − 1)kg(E1 , E1 ) = (n − 1)k.
i=2
Hence, we have
n L
(n − 1)π 2
Z
X
2 π
I(Vi , Vi ) ≤ sin t − (n − 1)k dt < 0
i=2 0 L L2
since we assumed L > √πk . At least one of the Vi ’s gives I(Vi , Vi ) < 0, and this shows γ
cannot be a minimizing geodesic. It leads to a contradiction.
√
(M, g) is compact because expp now maps B(0, π/ k) onto M . To argue that it has
finite fundamental group, we consider its universal cover π : M f → M . Recall that M
f
∗ ∗
admits a Riemannian metric π g which is locally isometric to g, so (M , π g) also satisfies
f
the same Ricci curvature lower bound. This shows M f is compact too, and consequently
π −1 (p) is a finite set for any p ∈ M . This shows π1 (M ) is a finite group.
Remark 11.11. The Ricci condition cannot be relaxed to Ric > 0. Any non-compact
regular surface in R3 with positive Gauss curvature serves as a counter-example, and there
are plenty of them! Recall that for regular surfaces, we have K = 2R and Ric = R2 g.
Remark 11.12. As the circle S1 has an infinite π1 (isomorphic to Z) and so is S1 × M for
any complete Riemannian manifold M , by the Bonnet-Myers’ Theorem it is impossible
for S1 × M to admit a Riemannian metric whose Ricci curvature has a uniform positive
lower bound. If M is compact, it is even impossible for S1 × M to admit a Riemannian
metric whose positive Ricci curvature.
The equality of Bonnet-Myers’ Theorem was proved by Cheng:
Theorem 11.14 (Synge). Let (M, g) be a compact Riemannian manifold with positive
sectional curvature. Then,
• if dim M is even and M orientable, then M is simply-connected;
• if dim M is odd, then it must be orientable.
278 11. Curvatures and Topology
The proof of the theorem is based on two lemmas, one analytic and another linear
algebraic.
Lemma 11.15. Let (M, g) be a compact Riemannian manifold, then in every free homo-
topy class [γ] of smooth closed curves , there exists a closed smooth geodesic γ
e (smooth at
based point too) that minimizes the arc-length among all curves in [γ].
Lemma 11.16. Any orthogonal matrix A ∈ O(n) with det(A) = (−1)n−1 must have 1
as one of its eigenvalues.
Proof. Any real eigenvalue of A ∈ O(n) is either 1 or −1. It can be shown by considering
Av = λv, so that
2 2 2
kvk = v T AT Av = (Av)T (Av) = Av · Av = kAvk = λ2 kvk
If n is even, then det(A) = −1. Since complex eigenvalues occur as conjugate pairs,
the product of all complex (non-real) eigenvalues is positive, and hence the product
of real eigenvalues must be negative. There are even many real eigenvalues counting
multiplicity, so at least one of the real eigenvalue is 1. If n is odd, then det(A) = 1. There
are odd many real eigenvalues counting multiplicity. We have at least one of the real
eigenvalue must be 1.
Proof of Theorem 11.14. We first prove the first statement. Suppose dim M is even and
M is orientable, but M is not simply-connected. Take a non-trivial homotopy class [γ]
of closed curves with γ being the minimizer of arc-length within the class (such γ exists
thanks to Lemma 11.15). Consider the parallel transport map Pγ : Tγ(0) M → Tγ(0) M
along γ which has determinant 1 by orientability. Note that Pγ (γ̇(0)) = γ̇(0) as γ is a
smooth closed geodesic. Consider the orthogonal complement E of span{γ̇(0)} in Tγ̇(0) M
so that E is invariant under Pγ and det(Pγ E ) = 1. Note that E has odd dimension, by
Lemma 11.16, there exists an eigenvector X0 ∈ E such that
Pγ (X0 ) = X0 .
Extend X0 by parallel transport along γ so that ∇γ̇ X(t) = 0 and X(0) = X0 , then we
get: Z
2
I(X, X) = |∇γ̇ X| − Rm(X, γ̇, γ̇, X) < 0,
γ | {z } | {z }
=0 >0
1
hence γ is not a minimizing geodesic . It leads to a contradiction, completing the proof
of the first statement.
The second statement can be proved in a similar way. Suppose dim M is odd, but M
is not orientable. Then, one can find a closed curve γe such that det Pγe = −1. Let γ ∈ [e γ]
be a smooth closed minimizing geodesic in the free homotopy class [e γ ], then we still have
det Pγ = −1 by continuity. Since Pγ (γ̇(0)) = γ̇(0), and the orthogonal complement E of
span{γ̇(0)} has odd dimension. By Lemma 11.16, there exists an eigenvector X0 ∈ E
such that Pγ (X0 ) = X0 . The rest of the proof goes exactly as in the even dimension case.
1Note that the index form we have discussed before require the variation vector field to vanish at end points, which is not
the case here. However, as the curve γ and vector field X(t) is smooth everywhere including the base point, the boundary
terms of integration by parts also vanish. Therefore, the second variation formula of L(γ) along X is also given by the index
form I(X, X).
11.3. Spaces of Constant Sectional Curvatures 279
Proof. One key ingredient of the proof is that the existence of a Jacobi field V (t) along
a geodesic γ(t) : [a, b] → M with V (a) = V (b) = 0 is equivalent to the non-invertibility
of (expγ(a) )∗ at the point corresponding to γ(b) (see Proposition 11.5). On the other
hand, such a Jacobi field does not exist if the metric has non-positive sectional curvature
according to (11.2). Hence, (expp )∗ is always invertible, and by the inverse function
theorem, it is a local diffeomorphism everywhere on Tp M . In particular, ge := (expp )∗ g
defines a Riemannian metric on Tp M , and (Tp M, ge) and (M, g) are locally isometric
through the map expp .
We first show that expp is surjective. Given a point q ∈ M , we let γ be the minimizing
unit-speed geodesic from p to q. Suppose d(p, q) = r > 0, then the geodesic γ is given by
γ(t) = expp (tγ̇(0)).
Hence, we have q = γ(r) = expp (rγ̇(0)). This shows expp is surjective.
Next, we show that for each q ∈ M , there exists ε > 0 such that
{Bε (Q)}Q∈expp−1 (q)
is a disjoint collection of open sets in Tp M . We pick such an ε > 0 so that Bε (q) is a
geodesic ball such that all geodesics from q leaves the ball through ∂Bε (q) (before they
come back to the ball, if ever). Index the set exp−1 p (q) by {Qα }. For any distinct pair
of Qα and Qβ , we connect them through a minimizing geodesic γ e (with respect to the
metric ge). Then, the curve γ := expp ◦eγ is a geodesic on M from q to q. However, by our
choice of ε, such a geodesic γ must go outside the geodesic ball Bε (q), and hence has
length > 2ε. This shows Qα and Qβ must be more than 2ε apart, so Bε (Qα ) and Bε (qβ )
are disjoint.
Then we argue that
G
exp−1
p Bε (q) = Bε (Qα ).
α
The ⊃-part is easy. Suppose X ∈ Bε (Qα ) for some Qα . As expp is a local isometry and
d(X, Qα ) < ε, so we have d expp (X), expp (Qα )) < ε too. It shows X ∈ exp−1
p Bε (q) .
280 11. Curvatures and Topology
Conversely, if Y ∈ exp−1
p Bε (q) , then consider the points y := expp (Y ) and q in
M . Let γ be a minimizing geodesic from y to q, and γ e be the geodesic lifted on Tp M ,
i.e. γ
e is geodesic on (Tp M, ge) such that γ e˙ (0)) = γ̇(0). Let r be
e(0) = Y , and (expp )∗ (γ
the length of γ so that d(y, q) = r < ε, then we have expp γ e(r) = γ(r) = q. Hence
Q := γ e(r) ∈ exp−1
p (q). Note that d(Q, Y ) = d(q, y) = r < ε by isometry, so Y ∈ Bε (Q).
This proves ⊂-part of our claim. It completes the proof that expp is a covering map.
11.3.2. Gauss Lemma. We next prove a fundamental result about radial tangent
vectors from 0 in Tp M and its image curve under the exponential map.
∂
Proof. Consider a curve σ(s) ⊂ ∂Br (0) ⊂ Tp M such that σ(0) = r ∂r and σ 0 (0) = X,
and define a family of geodesics
For each fixed s, the curve γs (t) is a geodesic by the definition of expp . Next we compute
∂ d
γs (t) = expp (0 · σ(s)) = 0
∂s (s,t)=(0,0) ds s=0
∂ d
γs (t) = expp (tσ(0))
∂t (s,t)=(0,0) dt t=0
= (expp )∗0 (σ(0)) = σ(0)
∂ d
γs (t) = expp (σ(s))
∂s (s,t)=(0,1) ds s=0
= (expp )∗σ(0) (σ 0 (0))
= (expp )∗r∂r (X)
∂ d
γs (t) = expp (tσ(0))
∂t (s,t)=(0,1) dt t=1
d
= expp (σ(0) + tσ(0))
dt t=0
= (expp )∗σ(0) (σ(0))
= (expp )∗r∂r (r∂r)
In short, the Gauss’s Lemma asserts that radial lines and round spheres in Tp M
remain to be orthogonal under the image of expp . We will use this lemma in the next
subsection to classify spaces with constant sectional curvatures.
To find u0 (0), we consider the results from Exercise 11.2, which shows
Xi ∂
V (t) = t
r ∂ui γ(t)
d Xi ∂ |X|g
u0 (0) = |V (t)| = = .
dt t=0 r ∂ui γ(0) r
Under geodesic normal coordinates at p, we also have |X|g = kXk, where k·k is the
standard Euclidean norm of Tp M ∼
= Rn . Therefore, we conclude that |V (r)|g depends
only on r, C, and kXk.
By exactly the same argument, we can get the same result for (g expp )∗ (X) ge. There-
fore, we have:
kXk √
r√C sin(r C)
if C > 0
expp )∗r∂r (X) ge = (expp )∗r∂r (X) g = |V (r)| = kXk
(g if C = 0
kXk
√
√
r −C
sinh(r −C) if C < 0
for any X ∈ Tr∂r (∂Br (0)).
Now given an arbitrary X ∈ Tr∂r (Tp M ), one can decompose it into
X = Xrad + Xsph
where Xrad is the radial component, and Xsph is tangential to ∂Br (0). By Gauss’s Lemma,
we have
g (expp )∗r∂r (Xrad ), (expp )∗r∂r (Xsph ) = 0.
11.3. Spaces of Constant Sectional Curvatures 283
for any r > 0, X ∈ Tr∂r (Tp M ) in the case of C = 0 and C < 0. Hence, g and ge are
isometric in these cases.
Finally, we deal with the case C > 0. Let (Sn , ge) be the round sphere with sectional
curvature C. With the same notations as the above, one can also argument in the same
way that g is isometric to ge locally (in the region on which expp is a diffeomorphism).
Therefore, by compactness of M (guaranteed by Bonnet-Myers’ Theorem), one can cover
M by finitely many open geodesic balls {Bα } each of which is isometric to another
geodesic ball {Beα } on Sn via the map, say, ϕα : Bα ⊂ M → B eα ⊂ Sn .
We want to glue these local isometries ϕα ’s to form a global isometry. However, it is
not a priori true that any pair ϕα and ϕβ of local isometries must agree on the overlap.
However, by the transitive action of SO(n) acting on Sn , one can compose an isometry
Φαβ : Sn → Sn that maps ϕα (Bα ∩ Bβ ) to ϕβ (Bα ∩ Bβ ) isometrically and Φαβ ◦ ϕα agrees
with ϕβ on the overlap. Replace ϕα by Φαβ ◦ ϕα . Repeat this replacement process for
each overlap (there are finitely many), one can construct a global isometry ϕ : M → Sn .
It completes the proof of the case C > 0.
[dC76] Manfredo P. do Carmo, Differential geometry of curves and surfaces, Prentice-Hall, Inc.,
Englewood Cliffs, N.J., 1976, Translated from the Portuguese. MR 0394451
[dC94] , Differential forms and applications, Universitext, Springer-Verlag, Berlin, 1994,
Translated from the 1971 Portuguese original. MR 1301070
[Küh05] Wolfgang Kühnel, Differential geometry, curves-surfaces-manifolds, Student Mathematical
Library, vol. 16, American Mathematical Society, Providence, RI, 2005. MR 3443721
[Lee09] Jeffrey M. Lee, Manifolds and differential geometry, Graduate Studies in Mathematics,
vol. 107, American Mathematical Society, Providence, RI, 2009. MR 2572292
[Lee13] John M. Lee, Introduction to smooth manifolds, second ed., Graduate Texts in Mathemat-
ics, vol. 218, Springer, New York, 2013. MR 2954043
[Mun00] James Munkres, Topology, 2 ed., Prentice Hall, 2000.
285