Multivariate Calculus 2011 TOCand Chapter 1
Multivariate Calculus 2011 TOCand Chapter 1
James S. Cook
Liberty University
Department of Mathematics
Fall 2011
2
preface
I do use the textbook, however, I follow these notes. You should use both. From past experience
I can tell you that the students who excelled in my course were those students who both studied
my notes and read the text. They also came to every class and paid attention. I recommend the
following course of study:
1. submit yourself to learn, keep a positive attitude. This course is a lot of work. Yes, probably
more than 3 others for most people. Most people have a lot of work to do in getting up to
speed on real mathematical thinking. There is no substitute for time and effort. If you’re
complaining in your mind about the workload etc... then you’re wasting your time.
4. attempt Problem Sets, you will likely find forming a study group is essential for success here.
I ask some very hard questions. The majority of the hwk grade comes from Problem Sets.
format of my notes
These notes were prepared with LATEX. You’ll notice a number of standard conventions in my notes:
5. often figures in these notes were prepared with Graph, a simple and free math graphing
program, or Maple, or Mathematica. Or some online math tool, of which there are dozens.
By now the abbreviations below should be old news, but to be safe I replicate them here once more:
3
Finally, please be warned these notes are a work in progress. I look forward to your input on how
they can be improved, corrected and supplemented. This is my second set of notes for calculus III.
I follow approximately the same LATEXformat as I used for the Calculus I and II notes of 2010-2011.
The old handwritten calculus III notes are not bad, but I want to improve the theoretical aspects
of these notes. I have included most of the examples from the old notes and also a myriad of
homework solutions in this version. There are far too many examples for lecture. My goal was to
gather the examples in a convenient format for the student.
Please understand that the definitions given in these notes are primary. I recommend
Stewart for additional examples, but not for definitions in this course. These notes take
a different (and I hope clearer) path through the subject matter of multivariable calculus, I make no
effort to be consistent with the text’s development of concepts and I intend to add much detail on
how calculus III works in curvelinear coordinates (not in Stewart). It is my intent these notes are a
self-contained treatment of differential multivariate calculus. However, you should understand that
I do assume you have a complete and working knowledge of Calculus I and II, basically Chapters
4
1-12 of Stewart, or take a look at my website for a breakdown of Math 131-132 as offered at Liberty
University.
At this juncture there are about 285 pages. I would like to add another 50 pictures, but, I’ll save
them for class. These notes only cover the material for the first two tests. I will probably transition
to the old notes after that point. However, I have more work to do on the last half of the course so
I would kindly ask you refrain from printing the old notes on integration and vector calculus. It is
highly likely that I am going to edit them and add discussion, proofs and totally new calculations.
I will let you know when I have notes beyond those found here1
The old notes and much more can be found at my calculus III webpage. Every so often I mention
the existence of an animation on my webpage. I am opening a zoo of gif-files where you can go see
all sorts of mathematical creatures. If you create interesting creatures I will (with your permission)
happily add them to my collection.
1
of course, if you’ve got money to burn and/or you just want to offend Al Gore by wasteful printing then go for
it, I always respect those who wish to make Al Gore cry.
Contents
1 analytic geometry 9
1.1 vectors in euclidean space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
1.2 the cross product . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29
1.3 lines and planes in R 3 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
1.3.1 parametrized lines and planes . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
1.3.2 lines and planes as solution sets . . . . . . . . . . . . . . . . . . . . . . . . . . 42
1.3.3 lines and planes as graphs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46
1.3.4 on projections onto a plane . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47
1.3.5 additional examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50
1.4 curves . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
1.4.1 curves in two-dimensional space . . . . . . . . . . . . . . . . . . . . . . . . . . 54
1.4.2 how can we find the Cartesian form for a given parametric curve? . . . . . . 62
1.4.3 curves in three dimensional space . . . . . . . . . . . . . . . . . . . . . . . . . 67
1.5 surfaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
1.5.1 surfaces as graphs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
1.5.2 parametrized surfaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72
1.5.3 surfaces as level sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75
1.5.4 combined concept examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
1.6 curvelinear coordinates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
1.6.1 polar coordinates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
1.6.2 cylindrical coordinates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84
1.6.3 spherical coordinates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87
5
6 CONTENTS
4 differentiation 149
4.1 directional derivatives . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 150
4.2 partial differentiation in R 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 153
4.2.1 directional derivatives and the gradient in R 2 . . . . . . . . . . . . . . . . . . 161
4.2.2 gradient vector fields . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 167
4.2.3 contour plots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 171
4.3 partial differentiation in R 3 and Rn . . . . . . . . . . . . . . . . . . . . . . . . . . . . 175
4.3.1 directional derivatives and the gradient in R 3 and Rn . . . . . . . . . . . . . 179
4.3.2 gradient vector fields in R 3 and Rn . . . . . . . . . . . . . . . . . . . . . . . . 183
4.4 the general derivative . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 188
4.4.1 matrix of the derivative . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 188
4.4.2 tangent space as graph of linearization . . . . . . . . . . . . . . . . . . . . . . 190
4.4.3 existence and connections to directional differentiation . . . . . . . . . . . . . 192
4.4.4 properties of the derivative . . . . . . . . . . . . . . . . . . . . . . . . . . . . 200
4.5 chain rules . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 203
4.6 tangent spaces and the normal vector field . . . . . . . . . . . . . . . . . . . . . . . . 218
4.6.1 level surfaces and tangent space . . . . . . . . . . . . . . . . . . . . . . . . . . 219
4.6.2 parametrized surfaces and tangent space . . . . . . . . . . . . . . . . . . . . . 220
4.6.3 tangent plane to a graph . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 223
4.7 partial differentiation with side conditions . . . . . . . . . . . . . . . . . . . . . . . . 226
4.8 gradients in curvelinear coordinates . . . . . . . . . . . . . . . . . . . . . . . . . . . . 238
4.8.1 polar coordinates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 238
4.8.2 cylindrical coordinates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 239
4.8.3 spherical coordinates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 239
5 optimization 243
5.1 lagrange multipliers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 244
5.1.1 proof of the method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 244
5.1.2 examples of the method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 245
5.1.3 extreme values of a quadratic form on a circle . . . . . . . . . . . . . . . . . . 258
5.1.4 quadratic forms in n-variables* . . . . . . . . . . . . . . . . . . . . . . . . . . 262
5.2 multivariate taylor series . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 263
5.2.1 taylor’s polynomial for one-variable . . . . . . . . . . . . . . . . . . . . . . . . 263
5.2.2 taylor’s multinomial for two-variables . . . . . . . . . . . . . . . . . . . . . . 264
5.2.3 taylor’s multinomial for many-variables . . . . . . . . . . . . . . . . . . . . . 267
CONTENTS 7
6 integration 285
analytic geometry
Euclidean space is named for the ancient mathematician Euclid. In Euclid’s view geometry was
a formal system with axioms and constructions. For example, the fact two parallel lines never
interect is called the parallel postulate. If you take the course in modern geometry then you’ll study
more of the history of Euclid. Fortunately for us the axiomatic/constructive approach was replaced
by a far easier view about 400 years ago. Rene Descartes made popular the idea of using numbers
to label points in space. In this new Cartesian geometry the fact two parallel lines in a plane never
intersect could be checked by some simple algebraic calculation. More than this, all sorts of curves,
surfaces and even more fantastic objects became constructible in terms of a few simple algebraic
equations. The idea of using numbers to label points and the resulting geometry governed by the
analysis of those numbers is called analytic geometry. We study the basics of analytic geometry
in this chapter.
In your previous mathematical studies you have already dealt with analytic geometry in the plane.
Trigonometry provided a natural framework to decipher all sorts of interesting facts about trian-
gles. Moreover, the study of trigonometric functions in turn has allowed us solutions to otherwise
intractable integrals in calculus II. Trigonometric substitution is an example of where the geometry
of triangles has allowed deeper analysis. It goes both ways. Geometry inspires analysis and analysis
unravels geometry. These are two sides of something deeper.
In this course we need to tackel three dimensional problems. The proper notation which groups
together concepts in the most efficient and clean manner is the vector notation. Historically, it was
predated by the quaternionic analysis of Hamilton, but for about 120 year the vector notation has
been the dominant framework for two and three dimensional analytic geometry1 . In particular,
the dot and cross products allow us to test for how parallel two lines are, or to project a line onto
a plane, or even to calculate the direction which is perpendicular to a pair of given directions.
Engineering and basic everyday physics all written in this vector langauge.
1
general geometries are more naturally understood in the language of differential forms and manifolds, but this is
where we all must begin
9
10 CHAPTER 1. ANALYTIC GEOMETRY
We also continue our study of functions in this chapter. We have studied functions f : U ⊆ R → R
in the first two semesters of calculus. One goal in this course is to extend the analysis to functions
of many variables. For example, f : R 2 → R with f (x, y) = x2 + y 2 . What can we say about
this function? What calculus is known to analyze the properties of f ? Before we begin to answer
such questions in future chapters, we need to spend some time on the basic geometry of Rn and
then especially R 3 . Naturally, we use R 3 to model the three spatial dimensions which frame our
everyday existence2 .
In this chapter we are concerned with understanding how to analytically describe points, curves
and surfaces. We will examine what the solution set of z = f (x, y) looks like for various f . Or,
what is the solution set of F (x, y) = k, or F (x, y, z) = k? We learn how think about mappings
t 7→ hx(t), y(t), z(t)i or (u, v) 7→ hx(u, v), y(u, v), z(u, v)i. What is the geometry of such mappings?
These are questions we hope to answer, at least in part, in this chapter.
Definition 1.1.1. .
R 2 = R × R = {(x, y) | x, y ∈ R}
R 3 = R × R × R = {(x, y, z) | x, y, z ∈ R}
Rn = |R × R ×
{z· · · × R} = {(x1 , x2 , . . . , xn ) | xj ∈ R f or each j ∈ Nn }
n copies
The fact that the n-tuples above are ordered means that two n-tuples are equal iff each and every
entry in the n-tuple matches.
2
I would not say we live in R 3 , it’s just a model, it’s not reality. Respectable philosophers as recently as 200
years ago labored under the delusion that euclidean space must be reality since that was all they could imagine as
reasonable
1.1. VECTORS IN EUCLIDEAN SPACE 11
The sometimes the term euclidean is added to emphasize that we suppose distance between points
is measured in the usual manner. Recall that in the one-dimensional p case the distance between
x, y ∈ R is given by the absolute value function; d(x, y) = |y − x| = (y − x)2 . We define distance
in n-dimensions by similar formulas:
It is simple to verify that the definition above squares with our traditional ideas about distance from
previous math courses. In particular, notice these follow from the Pythagorean theorem applied to
appropriate triangles. The picture below shows the three dimensional distance formula is consistent
with the two dimensional formula.
12 CHAPTER 1. ANALYTIC GEOMETRY
Notice that there is a natural correspondance between points and directed line-segments from the
origin. We can view an n-tuple p as either representing the point (p1 , p2 , . . . , pn ) or the directed
line-segment from the origin (0, 0, . . . , 0) to the point (p1 , p2 , . . . , pn ).
We will use the notation p~ for n-tuples throughout the remainder of these notes to emphasize the
fact that p~ is a vector. Some texts use bold to denote vectors, but I prefer the over-arrow notation
which is easily duplicated in hand-written work. The directed line-segment from point P to point
Q is naturally identified with the vector P~ − Q~ as illustrated below:
1.1. VECTORS IN EUCLIDEAN SPACE 13
This is consistent with the identification of points and vectors based at the origin. See how the
vector ~a and ~b are connected by the vector ~b − ~a
In the diagram below we illustrate the geometry behind the vector equation
~ =A
R ~+B
~ +C
~ + D.
~
Continuiing in this way we can add any finite number of vectors in the same tip-2-tail fashion.
1.1. VECTORS IN EUCLIDEAN SPACE 15
One important point to notice here is that we can naturally move vectors from the origin to other
points. Moreover, some people define vectors from the outset as directed-line-segments between
−−→ −−→
points. In particular, the vector from A to B in Rn is denoted AB and is defined by AB = B − A.
−−→
It is natural to suppose the vector AB is based at A, however, we can equally well picture the
−−→
vector AB based at any point in Rn .
If we wish to keep track of the base point of vectors then additional notation is required. We could
~ ) denotes a vector V
say that (p, V ~ which has basepoint p. Then addition, scalar multiplication,
dot-products and vector lengths are all naturally defined for such objects. You just do those op-
erations to the vector V~ and the point rides along. We will not use this notation in this course.
Instead, we will use words or pictures to indicate where a given vector is based. Sometimes vectors
are based at the origin, sometimes not. Sorry if this is confusing, but this is the custom of almost
all authors and if I invent notation and am more careful on this point then I’m afraid I may put
you at a disadvantage in other courses.
Algebraically, vector addition and scalar multiplication are easier to summarize concisely:
in R 2 ,
hx1 , y1 i + hx2 , y2 i = hx1 + x2 , y1 + y2 i
chx1 , y1 i = hcx1 , cy1 i
Or for R 3 ,
hx1 , y1 , z1 i + hx2 , y2 , z2 i = hx1 + x2 , y1 + y2 , z1 + z2 i
chx1 , y1 , z1 i = hcx1 , cy1 , cz1 i.
Generally, for ~x, ~y ∈ Rn and c ∈ R we define ~x + ~y and c~x component-wise as follows:
( ~x + ~y )j = ~xj + ~yj
( c~x )j = c~xj
for j = 1, 2, . . . , n.
16 CHAPTER 1. ANALYTIC GEOMETRY
Given these definitions it is often convenient to break a vector down into its vector components.
In particular, for R 2 , define x
b = h1, 0i and yb = h0, 1i hence:
The vector component of ha, bi in the x-direction3 is simply a x b whereas the vector
component of ha, bi in the y-direction is simply b yb. In contrast, a, b are the scalar com-
ponents of ha, bi in the x, y-directions respective.
Scalar components are scalars whereas vector components are vectors. These are entirely different
objects if n 6= 1, please keep clear this distinction in your mind. Notice that the vector components
are what we use to reproduce a given vector by the tip-to-tail sum:
The vector components of ha, b, ci are: a x b in the x-direction4 , b yb in the y-direction and
c zb in the z-direction. In contrast, a, b, c are the scalar components of ha, b, ci in the
x, y, z-directions respective.
Again, I emphasize that vector components are vectors whereas components or scalar components
are by default scalars.
1.1. VECTORS IN EUCLIDEAN SPACE 17
The story for Rn is not much different. Define for j = 1, 2, . . . , n the vector x
bj = h0, 0, . . . , 1, . . . , 0i
where the 1 appears in the j-th entry, hence:
If ~v = ha1 , a2 , . . . , an i then aj x
bj is the vector component in the xj -direction of ~v whereas
aj is the scalar component in the xj -direction of ~v .
Here’s an attempt at the picture for n > 3 (I use the linear algebra notation of e1 = x
b etc...):
18 CHAPTER 1. ANALYTIC GEOMETRY
Trigonometry is often useful in applied problems. It is not uncommon to be faced with vectors
which are described by a length and a direction in the plane. In such a case we need to rely on
trigonometry to break-down the vector into it’s Cartesian components.
hV1 , V2 i • hW1 , W2 i = V1 W1 + V2 W2 .
In R 3 we define,
hV1 , V2 , V3 i • hW1 , W2 , W3 i = V1 W1 + V2 W2 + V3 W3 .
~ = Pn Vj x
In Rn we define, for V bj and W ~ = Pn Wj x bj
j=1 j=1
~ •W
V ~ = V 1 W1 + V 2 W2 + · · · + V n Wn .
It is important to notice that the dot-product takes in two vectors and outputs a scalar. It has a
number of interesting properties which we will often use:
Example 1.1.9. .
1.1. VECTORS IN EUCLIDEAN SPACE 19
1. commutative: A ~ =B
~ •B ~
~ • A,
~ • (B
2. distributive: A ~ + C)
~ =A
~ •B
~ +A
~ • C,
~
~ + B)
3. distributive: (A ~ •C
~ =A
~ •C
~ +B
~ • C,
~
~ • (cB)
4. scalars factor out: A ~ = (cA) ~ = cA
~ •B ~
~ • B,
~ •A
5. non-negative: A ~≥0,
6. no null-vectors: A ~ = 0 iff A
~ •A ~ = 0.
Proof: The proof of these properties is simple if we use the right notation. Observe
n
X n
X
A ~ =
~ •B Aj B j = ~ • A.
B j Aj = B ~
j=1 j=1
~ + C)
Thus the dot-product is commutative. Next, note that (B ~ j = Bj + Cj hence,
n
X n
X n
X
~ • (B
A ~ + C)
~ = Aj (Bj + Cj ) = Aj Bj + ~ •B
Aj C j = A ~ +A
~ • C.
~
j=1 j=1 j=1
The proof of item (3.) actually follows from the commutative property and the right-distributive
property we just proved since
~ + B)
(A ~ =C
~ •C ~ • (A
~ + B)
~ =C ~+C
~ •A ~ =A
~ •B ~ +B
~ •C ~
~ • C.
The proof of (4.) is left to the reader. Continue to (5.), note that
n
X
~ •A
A ~= Aj Aj = A21 + A22 + · · · + A2n
j=1
hence it is clear that A ~ is the sum of squares of real numbers and consequently A
~ •A ~ ≥ 0.
~ •A
~ ~ ~
Moreover, if A • A = 0 and A 6= 0 then there must exist at least one component, say Aj 6= 0 hence
~ •A
A ~ ≥ A2 > 0 which is a contradiction. Therefore, (6.) follows.
j
~ is simply the distance from the origin to the point which the vector points.
The length of a vector A
In particular, we denote the length of the vector A~ by ||A||
~ and it’s clear from the formula in the
~ ~
proof for A • A that p
~ = A
||A|| ~ •A
~
this formula holds for Rn . Sometimes the length of the vector A ~ is also called the norm of A.
~
The norm also has interesting properties which are quite similar to those which are known for the
absolute value function on R (in fact, ||x|| = |x| for x ∈ R).
20 CHAPTER 1. ANALYTIC GEOMETRY
~ = |c|||A||,
1. absolute value of scalar factors out: ||cA|| ~
~ + B||
2. triangle inequality: ||A ~ ≤ ||A||
~ + ||B||,
~
~ • B|
3. Cauchy-Schwarz inequality: |A ~ ≤ ||A||
~ ||B||.
~
~ ≥0,
4. non-negative: ||A||
~ = 0 iff A
5. only zero vector has zero length: ||A|| ~ = 0.
As I introduced in the proof above5 , if a vector has a length of one then it is called a unit-vector.
~ = A and it follows
When it is convenient and unambiguous we use the notation ||A||
~ = AA.
A b
x
b• x
b = 1, yb • yb = 1, zb • zb = 1.
Example 1.1.13. .
(
1 i=j
In summary, we have that x bi • x bj = δij = . This is a very interesting formula. It shows
0 i 6= j
that set of vectors { x
b1 , x bn } are all of unit-length and distinct vectors are have dot-products
b2 , . . . , x
which are zero.
5
I happened to find this argument in Insel, Spence and Friedberg’s undergraduate linear algebra text.
22 CHAPTER 1. ANALYTIC GEOMETRY
Let’s pause to make a connection to the standard angle θ and the cartesian components.
~ = cos(θ) x ~ = ~ ~
Note that V b + sin(θ) yb and V V • x
b x b+ V • yb yb. It follows that:
~
cos(θ) = V • x
b and ~
sin(θ) = V • yb.
You could use these equations to define the standard angle in retrospect. Naturally, the decompo-
sition above equally well applies to Rn :
n
X
~ = hV
~ ~ ~ ~
V • x
b1 , V • x
b2 , . . . , V • bn i =
x V • x
bj x
bj .
j=1
6
a term which is studied in linear algebra
1.1. VECTORS IN EUCLIDEAN SPACE 23
The study of geometry involves lengths and angles of shapes. We have all the tools we need to
define the angle θ between nonzero vectors
~ B
Let A, ~ be nonzero vectors in Rn . We define the angle between A
~ and B
~ by
A ~
~ •B
−1
θ = cos .
~ ||B||
||A|| ~
~ B
Note nonzero vectors A, ~ have ||A||
~ = ~ =
6 0 and ||B|| 6 0 thus the Cauchy-Schwarz inequality |A~ • B|
~ ≤
~ ||B|| ~ ~
~ implies A • B ≤ 1. It follows that the argument of the inverse cosine is within its domain.
||A|| ~ ~
||A|| ||B||
Moreover, since the standard inverse cosine has range [0, π] it follows the angle which is given by
the formula above is the smallest angle between the vectors. Of course, if θ is the angle between
~ B
A, ~ then geometry clearly indicates 2π − θ is the angle on the other side of the θ vertex. I think
a picture helps:
The careful reader will question how I know the formula really recovers the idea of angle that we
have previously used in our studies of trigonometry. All I have really argued thus far is that the
formula for θ is reasonable. Examine the triangle formed by A,~ B~ and C ~ =B ~ − A.
~ Notice that
~ ~ ~ ~ ~
A + C = B. Picture A and B as adjacent sides to an angle θ̃ which has opposite side C. ~ Let the
~ B,
lengths of A, ~ C~ be A, B, C respective.
C 2 = A2 + B 2 − 2AB cos(θ̃).
Thus, we find the generalization of angle for Rn agrees with the two-dimensional concept we’ve
explored in previous courses. Moreover, we discover a geometrically lucid formula for the dot-
product:
A ~ = ||A||
~ •B ~ ||B||
~ cos(θ)
~ = AA
or if we denote A ~ = BB
b and B b then
~ •B
A ~ = AB cos(θ).
The connection between this formula and the definition is nontrivial and is essentially equivalent
to the Law of Cosines. This means that this is a powerful formula which allows deep calculation of
geometrically non-obvious angles through the machinery of vectors. Notice:
~ B
If A, ~ are nonzero orthogonal vectors then the angle between them is π/2.
this observation is an immediate consequence of the the definition of orthogonal vectors and the
fact cos(π/2) = 0. We find that orthogonal vectors are in fact perpendicular (which is a known
term from geometry). In addition,
~ B
If A, ~ are parallel vectors then A
~ •B
~ = AB and θ = 0.
likewise,
~ B
If A, ~ are antiparallel vectors then A ~ = −AB and θ = π.
~ •B
The dot-product gives us a concrete method to test for whether two vectors point in the same
direction, opposite directions or are purely perpendiular.
8 ~ B
even in the context of Rn we can place A, ~ and B
~ −A
~ in a particular plane, this argument actually extends to
n-dimensions provided you accept the Law of Cosines is known in any plane
1.1. VECTORS IN EUCLIDEAN SPACE 25
Example 1.1.16. .
Example 1.1.17. .
We can do more than just measure angles. We can also use the dot-product to project vectors to
lines or even planes. In particular:
Definition 1.1.18. projection onto vector.
~ 6= 0 then P roj ~ (B)
~ = B~ •A ~ onto A.
~ We also
A A
b A b defines the vector projection of B
define the orthogonal complement of B ~ with respect to A
~ by Orth ~ (B)
~ =B~ − P roj ~ (B).
~
A A
We’ve already seen the projection formula implicitly in the formula V ~ = (V ~•xb) x ~ • yb) yb
b + (V
note that P roj xb(V~ ) = (V~•x b) x
b since the unit vector of the unit vector x b is just x
b. Likewise,
~ ~
P roj yb(V ) = (V yb) yb. Very well, you might not find this terribly interesting. However, if was to
•
ask where a perpendicular bisector of h2, 2, 1i intersects h3, 6, 9i then I doubt you could do it with
geometry alone. It’s simple with the projection formula:
P rojh2,2,1i h1, 2, 3i = h1, 2, 3i • h 32 , 23 , 13 i h 23 , 23 , 13 i
= 23 + 43 + 33 h 32 , 23 , 13 i
= h2, 2, 1i.
26 CHAPTER 1. ANALYTIC GEOMETRY
That was actually an accident. I’ll try to do an uglier example in lecture. We’ll use the projection
from time to time, it is a really nice tool to calculate things that are at times hard to picture
directly. Generically the projection can be pictured by:
Example 1.1.20. .
1.1. VECTORS IN EUCLIDEAN SPACE 27
The projections onto the coordinate planes are sometimes interesting. Clearly x
b, yb fit in the xy-
plane hence
P rojxy−plane (hv1 , v2 , v3 i) = ( x
b • hv1 , v2 , v3 i) x
b + ( yb • hv1 , v2 , v3 i) yb = hv1 , v2 , 0i.
Likewise, P rojzx−plane (hv1 , v2 , v3 i) = hv1 , 0, v3 i and P rojyz−plane (hv1 , v2 , v3 i) = h0, v2 , v3 i.
Example 1.1.22. .
Perhaps this material belongs with the larger discussion of planes. I included it here simply to
illustrate the utility of the dot-product.
28 CHAPTER 1. ANALYTIC GEOMETRY
Example 1.1.23. Judging the colinearity of two vectors is important to physics. The work done
by a force is maximized when the force is applied over a displacement which is precisely parallel
to the force. On the other hand, the work done by a perpendicular force is zero. The dot-product
captures all these concepts in a nice neat formula: the work W done by a constant force F~ applied
to an object undergoing a displacement 4~x is given by W = F~ • 4~x .
Much later in this course we turn to the question of calculating work done by nonconstant forces
over arbitrary curves.
Example 1.1.24. .
Example 1.1.25. If ~v is the velocity of a mass m then the kinetic energy is given by K = 21 m~v • ~v .
Example 1.1.26. Or, if ~v is the velocity of a mass m and F~ is the net-force on m then the power
developed by F~ is given by P = ~v • F~ .
1.2. THE CROSS PRODUCT 29
My intent in this section is to motivate the standard formula for this product and to prove some
of the standard properties of this cross product. These calculations are special to R 3 .
Remark 1.2.1.
Forbidden jitzu ahead, In Rn the story is a bit more involved, we can calculate the orthogonal
complement to span{A, ~ B}
~ and this produces an (n − 2)-dimensional space of orthogonal
~ ~
vectors to A, B. If n = 4 this means there is a whole plane of vectors which we could choose.
Only in the case n = 3 is the orthogonal complement simply a one-dimensional space, a
line.
Therefore, suppose A, ~ B ~ are nonzero, nonparallel vectors in R 3 . I’ll calculate conditions on A
~×B
~
which insure it is perpendicular to both A ~ and B.
~ Let’s denote A ~×B ~ = C.~ We should expect C
~
~ ~ ~ ~
is some function of the components of A and B. I’ll use A = hA1 , A2 , A3 i and B = hB1 , B2 , B3 i
whereas C ~ = hC1 , C2 , C3 i
0=C ~ •A~ = C1 A1 + C2 A2 + C3 A3
0=C ~ •B~ = C1 B 1 + C2 B 2 + C3 B 3
~ •A
Suppose A1 6= 0, then we may solve 0 = C ~ as follows,
A2 A3
C1 = − C2 − C3
A1 A1
~ •B
Suppose B1 6= 0, then we may solve 0 = C ~ as follows,
B2 B3
C1 = − C2 − C3
B1 B1
It follows, given the assumptions A1 6= 0 and B1 6= 0,
A2 A3 B2 B3
C2 + C3 = C2 + C3
A1 A1 B1 B1
30 CHAPTER 1. ANALYTIC GEOMETRY
Multiply by A1 B1 to obtain:
B1 A2 C2 + B1 A3 C3 = A1 B2 C2 + A1 B3 C3
Thus,
(A1 B2 − B1 A2 )C2 + (A1 B3 − B1 A3 )C3 = 0
One solution is simply C2 = A1 B3 − B1 A3 and C3 = A1 B2 − B1 A2 and it follows that C1 =
A2 B3 − B2 A3 . Of course, generally we could have vectors which are nonzero and yet have A1 = 0
or B1 = 0. The point of the calculation is not to provide a general derivation. Instead, my intent
is simply to show you how you might be led to make the following definition:
Definition 1.2.2. cross product.
~ B
Let A, ~ be vectors in R 3 . The vector A
~ ×B
~ is called the cross product of A
~ with B
~ and
is defined by
~×B
A ~ = h A2 B3 − A3 B2 , A3 B1 − A1 B3 , A1 B2 − A2 B1 i.
~ cross B
We say A ~ is A
~ × B.
~
Both of these identities should be utilized to check your calculation of a given cross product. Let’s
think about the formula for the cross product a bit more. We have
~×B
A ~ = (A2 B3 − A3 B2 ) x
b1 + (A3 B1 − A1 B3 ) x
b2 + (A1 B2 − A2 B1 ) x
b3
distributing,
~×B
A ~ = A2 B3 x
b1 − A3 B2 x b2 − A1 B3 x
b1 + A3 B1 x b3 − A2 B1 x
b2 + A1 B2 x b3
The pattern is clear. Each term has indices 1, 2, 3 without repeat and we can generate the signs
via the antisymmetric symbol ijk which is defined be zero if any indices are repeated and
With this convenient shorthand we find the nice formula for the cross product that follows:
3
X
~×B
A ~ = Ai Bj ijk x
bk
i,j,k=1
b × yb = zb,
x yb × zb = x
b, zb × x
b = yb
There are many popular mneumonics to remember these. The basic properties of the cross product
together with these formula allow us to quickly calculate some cross products (see Example 1.2.7 )
~ B,
Let A, ~ C
~ be vectors in R 3 and c ∈ R
~×B
1. anticommutative: A ~ = −B
~ × A,
~
~ × (B
2. distributive: A ~ + C)
~ =A
~×B
~ +A
~ × C,
~
~ + B)
3. distributive: (A ~ ×C
~ =A
~×C
~ +B
~ × C,
~
~ × (cB)
4. scalars factor out: A ~ = (cA)
~ ×B
~ = cA
~ × B,
~
Proof: once more, the proof is easy with the right notation. Begin with (1.),
3
X 3
X 3
X
~×B
A ~ = bk = −
Ai Bj ijk x bk = −
Ai Bj jik x ~ × A.
bk = −B
Bj Ai jik x ~
i,j,k=1 i,j,k=1 i,j,k=1
The key observation was that ijk = −jik for all i, j, k. If you don’t care for this argument then
you could also give the brute-force argument below:
~×B
A ~ = h A2 B3 − A3 B2 , A3 B1 − A1 B3 , A1 B2 − A2 B1 i
= −h A3 B2 − A2 B3 , A1 B3 − A3 B1 , A2 B1 − A1 B2 i
= −h B2 A3 − B3 A2 , B3 A1 − B1 A3 , B1 A2 − B2 A1 i
~ × A.
= −B ~ (1.4)
The proof of (3.) follows naturally from (1.) and (2.), note:
~ + B)
(A ~ ×C
~ = −C
~ × (A
~ + B)
~ = −C
~ ×A
~−C
~ ×B
~ =A
~×C
~ +B
~ × C.
~
The properties above basically say that the cross product behaves the same as the usual addition
and multiplication of numbers with the caveat that the order of factors matters. If we switch the
order then we must include a minus due to the anticommutivity of the cross product.
Example 1.2.4. .
There are a number of popular tricks to remember this rule. Let’s look at a particular example a
couple different ways:
Example 1.2.5. .
1.2. THE CROSS PRODUCT 33
Example 1.2.6. .
Therefore,
Technically, this formula is not really a determinant since genuine determinants are formed from
matrices filled with objects of the same type. In the hybrid expression above we actually have one
row of vectors and two rows of scalars. That said, I include it hear since many people use it and
I also have found it useful in past calculations. If nothing else at least it helps you learn what a
determinant is. That is a calculation which is worthwhile since determinants have application far
beyond mere cross products.
34 CHAPTER 1. ANALYTIC GEOMETRY
Example 1.2.7. .
The calculation above is probably not the quickest for the example at hand here, but it is faster
for other computations. For example, suppose A~ = h1, 2, 3i and B
~ = x
b then
~×B
A ~ = (x
b + 2 yb + 3 zb) × x
b = 2 yb × x
b + 3 zb × x
b = −2 zb + 3 yb.
There are a number of identities which connect the dot and cross products. These formulas require
considerable effort if you choose to use brute-force proof methods.
~ × (B
1. A ~ × C)
~ = (A
~ • C)
~ B~ − (A
~ • B)
~ C~
~ × (B
2. Jacobi Identity: A ~ × C)
~ +B
~ × (C
~ × A)
~ +C
~ × (A
~ × B)
~ = 0,
~ • (B
3. cyclicity of triple product: A ~ × C)
~ =B ~ • (C
~ × A)
~ =C ~ • (A
~ × B)
~
~ × B||
~ 2 = ||A||
~ 2 ||B||
~ 2− A ~ •B~ 2
4. Lagrange’s identity: ||A
Proof: I leave proof of (1.)Pand (2.) to the reader. Let’s see how (3.) is shown in the compact
~ × C)
notation. Note (B ~ k=
ij Bi Cj ijk hence
3
X 3
X
~ • (B
A ~ × C)
~ = Ak Bi Cj ijk
k=1 i,j=1
3
X
= Bi Cj Ak ijk
i,j,k=1
3
X
= Cj Ak Bi jki
i,j,k=1
3
X
= Ak Bi Cj kij
i,j,k=1
where we have used the cyclicity of the antisymmetric symbol ijk = jki = kij . The cyclicity of
the triple product follows. Try to prove this without summations, it takes considerable patience.
36 CHAPTER 1. ANALYTIC GEOMETRY
Now we turn our attention to Lagrange’s identity. I begin by quoting a useful identity connecting
the antisymmetric symbol and the Kroenecker delta,
3
X
ijk lmk = δil δjm − δim δjl
k=1
Consider that,
3
X
~ × B||
||A ~ 2= ~ × B)
(A ~ 2
k
k=1
X3 X
3 3
X
= Ai Bj ijk Al Bm lmk
k=1 i,j=1 l,m=1
3
X 3
X
= Ai Al Bj Bm ijk lmk
i,j,k,l,m=1 k=1
3
X
= Ai Al Bj Bm (δil δjm − δim δjl )
i,j,l,m=1
3
X 3
X
= Ai Al δil Bj Bm δjm − Ai Bm δim Bj Al δjl
i,j,l,m=1 i,j,l,m=1
3
X 3
X
= A2i Bj2 − Ai B i B j Aj
i,j=1 i,j=1
3
X 3
X 3
X 3
X
= A2i Bj2 − Ai B i B j Aj
i=1 i=1 i=1 j=1
~ 2 ||B||
~ 2− A ~ 2.
~ •B
= ||A||
~ •B
Use Lagrange’s identity together with A ~ = AB cos(θ),
~ × B||
||A ~ 2 = A2 B 2 − [AB cos(θ)]2 = A2 B 2 (1 − cos2 (θ)) = A2 B 2 sin2 (θ)
~×B
A ~ = AB sin(θ)n̂
1.2. THE CROSS PRODUCT 37
Example 1.2.10. .
Example 1.2.11. .
38 CHAPTER 1. ANALYTIC GEOMETRY
Example 1.2.12. .
Example 1.2.13. Another important application of the cross product to physics is the Lorentz
~ then the force due to
force law. If a charge q has velocity ~v and travels through a magnetic field B
the electromagnetic interaction between q and the field is F~ = q~v × B.
~
Finally, we should investigate how the dot and cross product give nice formulas for the area of a
~ B
parallelogram or the volume of a parallel piped. Suppose A, ~ give the sides of a parallelogram.
~×B
Area = || A ~ ||
~ B,
On the other hand, if A, ~ C
~ give the corner-edges of a parallelogram then
~ • (B
V olume = A ~ × C)
~
These formulas are connected by the following thought: the volume subtended by A, ~ B~ and the
~ ~ ~ B.
unit-vector n̂ from A × B = AB sin(θ)n̂ is equal to the area of the parallelogram with sides A, ~
Algebraically:
~ × B)
n̂ • (A ~ = n̂ • (AB sin(θ)n̂) = |AB sin(θ)| = ||A
~ × B||.
~
The picture below shows why the triple product formula is valid.
Example 1.2.14. .
Moreover, given this geometric interpretation we find a new proof (up to a sign) for the cyclic
property. By the symmetry of the edges it follows that | A ~ • (B
~ × C)
~ | = | B ~ • (C
~ × A)
~ | =
|C~ • (A
~ × B)
~ |. We should find the same volume no matter how we label width, depth and height.
40 CHAPTER 1. ANALYTIC GEOMETRY
9
in later sections we will refer to the solution set formulation as a level set formulation
10
I assume the reader has familarity with these terms from calculus II, although, I do provide general definitions
later in this chapter. These are the two most basic examples
1.3. LINES AND PLANES IN R 3 41
If you wish to select a subset of the line or plane above you can appropriately restrict the domain
of the parameters. For example, one is often asked to find the parametrization of a line-segment
from a point P to a point Q. I recommend the following approach: for 0 ≤ t ≤ 1 let
~r(t) = P (1 − t) + tQ.
It’s easy to calculate ~r(0) = P and ~r(1) = Q. This formula can also be written as
−−→
~r(t) = P + t(Q − P ) = P + t [P Q].
If we let t go beyond the unit-interval then we trace out the line which contains the line-segment P Q.
Example 1.3.3. Find the parametrization of a line segment which goes from (1, 3) to
(5, 2). We use the comment preceding this example and construct:
On the other hand, if we wish to parametrize just the parallelogram in the plane with corners ~ro ,
~ro + A,~ ~ro + B ~ and ~ro + A~+B~ we may limit the values of the parameters u, v to the unit square
[0, 1] × [0, 1]; that is, we demand 0 ≤ u ≤ 1 and 0 ≤ v ≤ 1.
42 CHAPTER 1. ANALYTIC GEOMETRY
If we allow (u, v) to trace out all of R 2 then we will find the parametrization above covers the
xy-plane. Scalar equations which capture the same are x = 1 + u, y = 2 + v, z = 0. If we restrict
the parameters to 0 ≤ u, v ≤ v then the mapping ~r just covers [1, 2] × [2, 3] × {0}.
Identify the point (1, 2, 3) is on the plane, in fact it ~r(0, 0) = ~ro . Moreover, the vectors h1, −1, 0i
and h1, 0, 1i lie on the plane. You can verify that h1, −1, 0i connects ~r(0, 0) = (1, 2, 3) and ~r(1, 0) =
(2, 1, 3) whereas h1, 1, 0i connects ~r(0, 0) = (1, 2, 3) and ~r(0, 1) = (2, 2, 4).
We say that S ⊂ R 3 is a plane with base point ~ro and normal vector ~n iff each ~r ∈ S
satisfies
(~r − ~ro ) • ~n = 0. (vector equation of plane)
The geometric motivation for this definition is simple enough: the normal vector is a vector which
is perpendicular to all vectors in the plane. If we take the difference ~r −~ro then this will be a vector
which lies on the plane and consequently we must insist they are orthogonal. Here’s a picture of
why the definition is reasonable:
Note that the same set of points S can be given many different base points and many different
normals. This reflects the fact that we can choose the base point anywhere on the plane and the
1.3. LINES AND PLANES IN R 3 43
normal either above or below the plane and can be given many different lengths.
Let ~r = hx, y, zi be an arbitrary point on the plane S with base point ~ro = hxo , yo , zo i and normal
~n = ha, b, ci then we can write our plane equation explicitly:
I expect you to know the vector and scalar equations for a plane. We will use both in concept and
calculation.
~ = x
Example 1.3.7. Suppose the plane S contains the point (1, 2, 3) and the vectors A b and
~
B = h1, 3, 4i. Find the scalar equation of S.
Solution: we are given the base point (1, 2, 3) so we need only find a normal for S. Recall that the
~ with B
cross product of A ~ gives a vector which is perpendicular to both vectors and by the geometry
of R 3 it must be colinear with the normal vector we seek. After all, one we use up two of the
dimensions then there is only one left to use in R 3 . Calculate:
~×B
~ = x
A b× x b× x
b + 3 yb + 4 zb = x b × yb + 4 x
b+ 3x b × zb = 3 zb − 4 yb.
−4(y − 2) + 3(z − 3) = 0.
a1 (x − xo ) + b1 (y − yo ) + c1 (z − zo ) = 0 and a2 (x − xo ) + b2 (y − yo ) + c2 (z − zo ) = 0
44 CHAPTER 1. ANALYTIC GEOMETRY
We must solve both at once to find an equation for L. Generally there is no simple formula, however
if a1 , b1 , c1 , a2 , b2 , c2 6= 0 then we are free to divide by those constants. First to get the equations
to match divide both by the coefficient of their (z − zo ) factor,
a1 b1 a2 b2
(x − xo ) + (y − yo ) + z − zo = 0 and (x − xo ) + (y − yo ) + z − zo = 0
c1 c1 c2 c2
Thus, solving both equations for z − zo we find,
a1 b1 a2 b2
(x − xo ) + (y − yo ) = (x − xo ) + (y − yo )
c1 c1 c2 c2
Multiply by c1 c2 and rearrange to find
a1 c2 − a2 c1 (x − xo ) + b1 c2 − b2 c1 (y − yo ) = 0
Consequently,
x − xo y − yo
= ?.
b1 c2 − b2 c1 a2 c1 − a1 c2
Following the same algebra we can equally well solve for x − xo ,
b1 c1 b2 c2
(y − yo ) + (z − zo ) = (y − yo ) + (z − zo )
a1 a1 a2 a2
and multiply by a1 a2 to find
b1 a2 − b2 a1 (y − yo ) + c1 a2 − c2 a1 (z − zo ) = 0
Hence,
y − yo z − zo
= ? ?.
c1 a2 − c2 a1 b2 a1 − b1 a2
We combine ? and ?? to obtain the symmetric equations for the line L
x − xo y − yo z − zo
= =
b1 c2 − b2 c1 c2 a1 − c1 a2 c2 a1 − c1 a2
If we denote ~n1 = ha1 , b1 , c1 i and ~n2 = ha2 , b2 , c2 i then recognize that
~ = ~n1 × ~n2 = hb1 c2 − b2 c1 , c2 a1 − c1 a2 , c2 a1 − c1 a2 i
w
~ = ha, b, ci, the symmetric equation is simply:
Therefore, with the notation w
x − xo y − yo z − zo
= = .
a b c
We can use these equations to parametrize the line L. Let t = x−x a
o
hence x = xo + at is the
parametric equation for x, likewise, y = yo + bt and z = zo + ct. We identify that ha, b, ci is
precisely the direction-vector for the line L since we can group the scalar parametric equations
above to obtatin the vector parametric equation below:
~r(t) = hxo + at, yo + bt, zo + cti = hxo , yo , zo i + tha, b, ci.
We find the following interesting geometric result:
1.3. LINES AND PLANES IN R 3 45
The line of intersection for two planes has a direction vector which is colinear
with the cross product of the normals of the intersected planes.
If we take a step back and analyze this by pure geometric visualization this is rediculously obvious.
The line of intersection lies in both planes. Therefore, if ~v is the direction vector of L and ~n1 is the
normal of plane S1 and ~n2 is the normal of plane S2 then
1. ~v • ~n1 = 0 because ~v lies on S1
3. if ~v is perpendicular to both ~n1 and ~n2 then it must be colinear with ~n1 × ~n2 .
This is an example of how geometry is sometimes easier than algebra. In fact, that is often the
case, however, you must get used to both lines of logic in this course. This is the beauty of analytic
geometry.
Let’s examine how we can get from the parametric viewpoint to the symmetric equations. Suppose
we are given the vector parametric equations for a line with base point ~ro = (xo , yo , zo ) and direction
vector ~v = ha, b, ci with a, b, c 6= 0:
Further suppose we denote ~r = hx, y, zi then the scalar parametric equations for the line are:
Finally, recall that I insisted that the intersection of the planes was a line from the outset of this
discussion. There is another possibility. It could be that two planes are either parallel and have
no point of intersection, or they could simply be the same plane. In both of those cases the cross
product of the normals is trivial since the normals of parallel planes are colinear.
clearly S = graph(f ) in this context. You could also write S as the graph y = g(x, z) or x = h(y, z)
provided b 6= 0 and a 6= 0 respective. I hope you can find the formulas for g or h. These graphs
provide parametrizations as follows, once more consider the case c 6= 0, let
x = u, y = v, z = f (u, v)
Equivalently,
~r(u, v) = hu, v, f (u, v)i.
In this way we find a natural parametrization of a graph. Likewise, if a 6= 0 or b 6= 0 then
~r(u, v) = hu, g(u, v), v)i or ~r(u, v) = hh(u, v), u, vi
provide natural parametrizations. Parametrizations created in these ways are said to be induced
from the graphs g, h respective.
Writing the line as a graph requires us to solve the symmetric equations for two of the cartesian
coordinates in terms of the remaining coordinate. For example, solve for y, z in terms of x:
x − xo y − yo z − zo
= = ⇒ a(y − yo ) = b(x − xo ) & a(z − zo ) = c(x − xo ).
a b c
1.3. LINES AND PLANES IN R 3 47
hence,
b c
y = yo + (x − xo ) & z = zo + (x − xo )
| a {z } | a {z }
let this be h(x) let this be g(x)
Define f (x) = (h(x), g(x)) then graph(f ) = {(x, f (x)) | x ∈ R} and it is clear that graph(f ) = L.
We say f : R → R 2 is a R 2 -valued function of a real variable. Some texts call such functions
mappings whereas they inist that functions are real-valued. I make no such restriction in these
notes. In any event, there is a natural parametrization which is induced from the graph for L, use
x = t hence11
~r(t) = ht, f (t)i = ht, g(t), h(t)i
parametrizes L. We could also solve for y or z provided b 6= 0 or c 6= 0. I leave those to the reader.
In three dimensions we can calculate this projection via subtracting off the piece of the vector which
is in the normal direction. If ~n is the normal to S then I claim
Let’s see why in n = 3 these formulas are equivalent. Construct the normal corresponding to the
unit vectors in the usual manner: ~n = u b2 . Orthogonality of the unit vectors implies θ = π2
b1 × u
hence ||b
u1 × u
b2 || = 1 and it follows ~n = n
b. Let us define
Notice that n
b•u
b1 = 0 and n
b•u
b2 = 0. Take the dot-product of ? with u
b1 and u
b2 to obtain,
P roj S (~v ) • u
b1 = ~v • u
b1 and P roj S (~v ) • u
b2 = ~v • u
b2
It follows that P roj S (~v ) = ~v • u b1 ub1 + ~v • u
b2 ub2 . The formulas agree. They both produce the
same projection onto the plane. If we attach ~v to the plane at some point P ∈ S then P rojS (~v )
attached to P will point to the point on the plane S which is closest to the end of the vector ~v .
Perhaps the example that follows will help you understand the discussion above.
Example 1.3.9. Let S be the plane through (0, 0, 1) with normal ~n = h1, 0, 1i. Notice that ~u1 =
h1, 0, −1i and ~u2 = h0, 1, 0i are orthogonal as ~u1 • ~u2 = 0 and they are both on S as ~n • ~u1 = 0 and
~n • ~u2 = 0. Normalize the vectors,
1 1
b1 = √ h1, 0, −1i,
u b2 = h0, 1, 0i
u b = √ h1, 0, 1i
n
2 2
Let ~v = ha, b, ci. Calculate,
P rojS (~v ) =~v • u
b1 ub1 + ~v • u
b2 ub2
1
= a − c h1, 0, −1i + bh0, 1, 0i
2
1 1
= h (a − c), b, (c − a) i.
2 2
1.3. LINES AND PLANES IN R 3 49
Example 1.3.10. .
50 CHAPTER 1. ANALYTIC GEOMETRY
Example 1.3.12. .
Example 1.3.13. .
Example 1.3.14. .
1.3. LINES AND PLANES IN R 3 51
Example 1.3.15. .
Example 1.3.16. .
52 CHAPTER 1. ANALYTIC GEOMETRY
Example 1.3.17. .
Example 1.3.18. .
1.3. LINES AND PLANES IN R 3 53
Example 1.3.19. .
Example 1.3.20. .
54 CHAPTER 1. ANALYTIC GEOMETRY
1.4 curves
A curve is a one-dimensional subset of some space. There are at least three common, but distinct,
ways to frame the mathematics of a curve. These viewpoints were already explored in the previous
section but I list them once more: we can describe a curve:
3. as a graph.
I expect you master all three viewpoints in the two-dimensional context. However, for three or more
dimensions we primarily use the parametric viewpoint in this course. Exceptions to this rule are
fairly rare: the occasional homework problem where you are asked to find the curve of intersection
for two surfaces, or the symmetric equations for a line. In contrast, the parametric decription of
a curve in three dimensions is far more natural. Do you want to describe a curve as where two
surfaces intersect or would you rather describe a curve as a set of points formed by pasting a copy
of the real line through your space? I much prefer the parametric view.
A vector valued function of a real variable is an assignment of a vector for each real number
in some domain. It’s a mapping t 7→ f~(t) = f1 (t), f2 (t), . . . , fn (t) for each t ∈ J ⊂ R.
We say fj : J ⊆ R → R is the j-th component function of f~. Let C = f~(J) then C is said
to be a curve which is parametrized by f~. We can also say that t 7→ f~(t) is a path in
Rn . Equivalently, but not so usefully, we can write the scalar parametric equations for C
above as
x1 = f1 (t), x2 = f2 (t), . . . , xn = fn (t)
for all t ∈ J.
When we define a parametrization of a curve it is important to give the formula for the path and
the domain of the parameter. Note that when I say the word curve I mean for us to think about
some set of points, whereas when I say the word path I mean to refer to the particular mapping
whose image is a curve. We may cover a particular curve with infinitely many different paths.
We know this is quite restrictive. We must satisfy the vertical line test if we say our curve is the
graph of a function.
Example 1.4.4. On the other hand, if we wish to describe the set of all points such that sin(y) = x
we also face a similar difficulty if we insist on functions having independent variable x. Naturally, if
we allow for functions with y as the independent variable then f (y) = sin(y) has graph graph(f ) =
{(f (y), y) | y ∈ dom(f )}. You might wonder, is this correct? I would say a better question is, ”is
this allowed?”. Different books are more or less imaginative about what is permissible as a function.
This much we can say, if a shape fails both the vertical and horizontal line tests then it is not the
graph of a single function of x or y.
Example 1.4.5. Let f (x) = mx + b for some constants m, b then y = f (x) is the line with slope
m and y-intercept b.
In an algebra class you might have called this the ”graph of an equation”, but that terminology
is dead to us now. For us, it is a level curve. Moreover, for a particular set of points C ⊆ R2
we can find more than one function F which produces C as a level set. Unlike functions, for a
particular curve there is not just one function which returns that curve. This means that it might
be important to give both the level-function F and the level k to specify a level curve F (x, y) = k.
Example 1.4.7. A circle of radius R centered at the origin is a level curve F (x, y) = R2 where
F (x, y) = x2 + y 2 . We call F the level function (of two variables).
Example 1.4.8. To describe sin(y) = x as a level curve we simply write sin(y)−x = 0 and identify
the level function is F (x, y) = sin(y) − x and in this case k = 0. Notice, we could just as well say
it is the level curve G(x, y) = 1 where G(x, y) = x − sin(y) + 1.
Note once more this type of ambiguity is one distinction of the level curve langauge, in constrast,
the graph graph(f ) of a function y = f (x) and the function f are interchangeable. Some mathe-
maticians insist the rule x 7→ f (x) defines a function whereas others insist that a function is a set
of pairs (x, f (x)). I prefer the mapping rule because it’s how I think about functions in general
whereas the idea of a graph is much less useful in general.
Example 1.4.9. A line with slope m and y-intercept b can be described by F (x, y) = mx+b−y = 0.
Alternatively, a line with x-intercept xo and y-intercept yo can be described as the level curve
G(x, y) = xxo + yyo = 1.
Example 1.4.10. Level curves need not be simple things. They can be lots of simple things glued
together in one grand equation:
Solutions to the equation above include the line y = x, the unit circle x2 +y 2 = 1, the tilted-hyperbola
known more commonly as the reciprocal function y = x1 and finally the graph of the tangent. Some
of these intersect, others are disconnected from each other.
It is sometimes helpful to use software to plot equations. However, we must be careful since they
are not as reliable as you might suppose. The example above is not too complicated but look what
happens with Graph:
1.4. CURVES 57
If y = f (x) is the graph of a function then we can write F (x, y) = f (x) − y = 0 hence the
graph y = f (x) is also a level curve.
Not much of a theorem. But, it’s true. The converse is not true without a lot of qualification. I’ll
state that theorem (it’s called the implicit function theorem) in a future chapter after we’ve studied
partial differentiation.
(x − h)2 (y − k)2
+ = 1.
a2 b2
(x − h)2 (y − k)2
− = 1.
a2 b2
Example 1.4.14. A spiral can be thought of as a sort of circle with a variable radius. With that
in mind I write: for t ≥ 0,
~r(t) = ht cos(t), t sin(t)i
Finding the parametric equations for a curve does require a certain amount of creativity. However,
it’s almost always some slight twist on the examples I give in this section. The remaining examples
I also give in calculus II, I add some detail to emphasize how the parametrization matches the
already known identities of certain curves and I add pictures which emphasize the idea that the
parametrization pastes a line into R 2 .
Example 1.4.15. Let x = R cos(t) and y = R sin(t) for t ∈ [0, 2π]. This is a parametrization of
the circle of radius R centered at the origin. We can check this by substituting the equations back
into our standard Cartesian equation for the circle:
Recall that cos2 (t) + sin2 (t) = 1 therefore, x(t)2 + y(t)2 = R2 for each t ∈ [0, 2π]. This shows that
the parametric equations do return the set of points which we call a circle of radius R. Moreover,
we can identify the parameter in this case as the standard angle from standard polar coordinates.
1.4. CURVES 59
Example 1.4.16. Let x = R cos(et ) and y = R sin(et ) for t ∈ R. We again cover the circle at t
varies since it is still true that (R cos(et ))2 + (R sin(et ))2 = R2 (cos2 (et ) + sin(et )) = R2 . However,
since range(et ) = [1, ∞) it is clear that we will actually wrap around the circle infinitly many times.
The parametrizations from this example and the last do cover the same set, but they are radically
different parametrizations: the last example winds around the circle just once whereas this example
winds around the circle ∞-ly many times.
Example 1.4.17. Let x = R cos(−t) and y = R sin(−t) for t ∈ [0, 2π]. This is a parametrization
of the circle of radius R centered at the origin. We can check this by substituting the equations back
into our standard Cartesian equation for the circle:
Recall that cos2 (−t)+sin2 (−t) = 1 therefore, x(t)2 +y(t)2 = R2 for each t ∈ [0, 2π]. This shows that
the parametric equations do return the set of points which we call a circle of radius R. Moreover,
we can identify the parameter an angle measured CW12 from the positive x-axis. In contrast, the
standard polar coordinate angle is measured CCW from the postive x-axis. Note that in this example
we cover the circle just once, but the direction of the curve is opposite that of Example 1.4.15.
12
CW is an abbreviation for ClockWise,whereas CCW is an abbreviation for CounterClockWise.
60 CHAPTER 1. ANALYTIC GEOMETRY
The idea of directionality is not at all evident from Cartesian equations for a curve. Given a graph
y = f (x) or a level-curve F (x, y) = k there is no intrinsic concept of direction ascribed to the curve.
For example, if I ask you whether x2 + y 2 = R2 goes CW or CCW then you ought not have an
answer. I suppose you could ad-hoc pick a direction, but it wouldn’t be natural. This means that
if we care about giving a direction to a curve we need the concept of the parametrized curve. We
can use the ordering of the real line to induce an ordering on the curve.
To reverse orientation for x = f (t), y = g(t) for t ∈ J = [a, b] one may simply replace t by
−t in the parametric equations, this gives new equations which will cover the same curve via
x = f (−t), y = g(−t) for t ∈ [−a, −b].
Example 1.4.20. The line-segment from (0, −1) to (1, 2) can be parametrized by x = t and y =
3t − 1 for 0 ≤ t ≤ 1. On the other hand, the line-segment from (1, 2) to (0, −1) is parametrized by
x = −t, y = −3t − 1 for −1 ≤ t ≤ 0.
The other method to graph parametric curves is simply to start plugging in values for the parameter
and assemble a table of values to plot. I have illustrated that in part by plotting the green dots in
the domain of the parameter together with their images on the curve. Those dots are the results of
plugging in the parameter to find corresponding values for x, y. I don’t find that is a very reliable
approach in the same way I find plugging in values to f (x) provides a very good plot of y = f (x).
That sort of brute-force approach is more appropriate for a CAS system. There are many excellent
tools for plotting parametric curves, hopefully I will have some posted on the course website. In
addition, the possibility of animation gives us an even more exciting method for visualization of the
62 CHAPTER 1. ANALYTIC GEOMETRY
time-evolution of a parametric curve. In the next chapter we connect the parametric viewpoint with
physics and such an animation actually represents the physical motion of some object. My focus
in the remainder of this chapter is almost purely algebraic, I could draw pictures to explain, but I
wanted the notes to show you that the pictures are not necessary when you understand the algebraic
process. That said, the best approach is to do some combination of algebraic manipulation/figuring
and graphical reasoning.
1. how can we find the Cartesian form for a given parametric curve?
In case (2.) we mean to include the ideas of level curves and graphs. It turns out that both
questions can be quite challenging for certain examples. However, in other cases, not so much: for
example any graph y = f (x) is easily recast as the set of parametric equations x = t and y = f (t)
for t ∈ dom(f ). For the standard graph of a function we use x as the parameter.
1.4.2 how can we find the Cartesian form for a given parametric curve?
Example 1.4.21. What curve has parametric equations x = t for y = t2 for t ∈ R? To find
Cartesian equation we eliminate the parameter (when possible)
t 2 = x2 = y ⇒ y = x2
Example 1.4.23. What curve has parametric equations x = t for y = t2 for t ∈ [0, 1]? To find
Cartesian equation we eliminate the parameter (when possible)
t 2 = x2 = y ⇒ y = x2
Thus the Cartesian form of the given parametrized curve is simply y = x2 , however, given that
0 ≤ t ≤ 1 and x = t it follows we do not have the whole parabola, instead just y = x2 for 0 ≤ x ≤ 1.
1.4. CURVES 63
Example 1.4.24. Identify what curve has parametric equations x = tan−1 (t) and y = tan−1 (t) for
t ∈ R. Recall that range(tan−1 (t)) = (−π/2, π/2). It follows that −π/2 < x < π/2. Naturally we
just equate inverse tangent to obtain tan−1 (t) = y = x. The curve is the open line-segment with
equation y = x for −π/2 < x < π/2. This is an interesting parameterization, notice that as t → ∞
we approach the point (π/2, π/2), but we never quite get there.
Example 1.4.25. Consider x = ln(t) and y = et − 1 for t ≥ 1. We can solve both for t to obtain
t = ex = ln(y + 1) ⇒ y = −1 + exp(exp(x)).
The domain for the expression above is revealed by analyzing x = ln(t) for t ≥ 1, the image of
[1, ∞) under natural log is precisely [0, ∞); ln[1, ∞) = [0, ∞).
(y − 3)2
cosh2 (t) − sinh2 (t) = 1 ⇒ (x + 1)2 − = 1.
4
Note that cosh(t) ≥ 1 hence x + 1 ≥ 1 thus x ≥ 0 for the curve described above. On the other
hand y is free to range over all of R since hyperbolic sine has range R. You should13 recognize the
equation as a hyperbola centered at (−1, 3).
Example 1.4.27. Find parametric equations for the circle centered at (h, k) with radius
R.
To begin recall the equation for such a circle is (x − h)2 + (y − k)2 = R2 . Our inspiration is the
identity cos( t) + sin2 (t) = 1. Let x − h = R cos(t) and y − k = R sin(t) thus
I invite the reader to verify these do indeed parametrize the circle by explicitly plugging in the
equations into the circle equation. Notice, if we want the whole circle then we simply choose any
interval for t of length 2π or longer. On the other hand, if you want to select just a part of the
circle you need to think about where sine and cosine are positive and negative. For example, if I
want to parametrize just the part of the circle for which x > h then I would choose t ∈ (− π2 , π2 ).
13
many students need to review these at this point, we use circles, ellipses and hyperbolas as examples in this
course. I’ll give examples of each in this chapter.
64 CHAPTER 1. ANALYTIC GEOMETRY
The reason I choose that intuitively is that the parametrization given for the circle above is basically
built from polar coordinates14 centered at (h, k). That said, to be sure about my choice of parameter
domain I like to actually plug in some of my proposed domain and make sure it matches the
desired criteria. I think about the graphs of sine and cosine to double check my logic. I know that
cos(− π2 , π2 ) = (0, 1] whereas sin(− π2 , π2 ) = (−1, 1), I see it in my mind. Then I think about the
parametric equations in view of those facts,
I see that x will range over (h, h + R] and y will range over (k − R, k + R). This is exactly what I
should expect geometrically for half of the circle. Visualize that x = h is a vertical line which cuts
our circle in half. These are the thoughts I think to make certain my creative leaps are correct. I
would encourage you to think about these matters. Don’t try to just memorize everything, it will
not work for you, there are simply too many cases. It’s actually way easier to just understand these
as a consequence of trigonometry, algebra and analytic geometry.
14
we will discuss further in a later section, but this should have been covered in at least your precalculus course.
1.4. CURVES 65
Example 1.4.28. Find parametric equations for the level curve x2 + 2x + 14 y 2 = 0 which
give the ellipse a CW orientation.
If we choose t ∈ [0, 2π) then the whole ellipse will be covered. I could simplify cos(−t) = cos(t) and
sin(−t) = − sin(t) but I have left the minus to emphasize the idea about reversing the orientation.
In the preceding example we gave the circle a CCW orientation.
Example 1.4.29. Find parametric equations for the part of the level curve x2 − y 2 = 1
which is found in the first quadrant.
Example 1.4.30. Find parametric equations for the part of the level curve x2 − y 2 = 1
which is found in the third quadrant.
Based on our thinking from the last example we just need to modify the solution a bit:
Note that if t ∈ (−∞, 0] then − cosh(t) ≤ −1 and sinh(t) ≤ 0, this puts us in the third quadrant. It
is also clear that these parametric equations to solve the hyperbola equation since
The examples thus far are rather specialized, and in general there is no method to find parametric
equations. This is why I said it is an art.
66 CHAPTER 1. ANALYTIC GEOMETRY
for t ≥ 2. These parametrizations simply cover different parts of the same level curve.
If you are at all like me when I first learned about parametric curves you’re probably
wondering what is t ? You probably, like me, suppose incorrectly that t should be just
like x or y. There is a crucial difference between x and y and t. The notations x and y
are actually shorthands for the Cartesian coordinate maps x : R2 → R and y : R2 → R
where x(a, b) = a and y(a, b) = b. When I use the notaion x = 3 then you know what
I mean, you know that I’m focusing on the vertical line with first coordinate 3. On the
other hand, if I say t = 3 and ask where is it? Then you should say, you question doesn’t
make sense. The concept of t is tied to the curve for which it is the parameter. There are
infinitely many geometric meanings for t. In other words, if you try to find t in the xy-
plane without regard to a curve then you’ll never find an answer. It’s a meaningless question.
On the other hand if we are given a curve and ask what the meaning of t is for that curve
then we ask a meaningful question. There are two popular meanings.
1. the parameter s measures the arclength from some base point on the given curve.
In case (1.) for an oriented curve this actually is uniquely specified if we have a starting
point. Such a parameterization is called the arclength parametrization or unit-speed
parametrization of a curve. These play a fundamental role in the study of the differential
geometry of curves. In case (2.) we have in mind that the curve represents the physical
trajectory of some object, as t increases, time goes on and the object moves. I tend to
use (2.) as my conceptual backdrop. But, keep in mind that these are just applications of
parametric curves. In general, the parameter need not be time or arclength. It might just
be what is suggested by algebraic convenience.
1.4. CURVES 67
Of course, we could also put a spiral on a plane through much the same device:
The idea of the last example can be used to create many interesting examples. These should suffice
for our purposes here. I really just want you to think about what a parametrization does. Moreover,
I hope you can find it in your heart to regard the parametric viewpoint as primary. Notice that any
curves in three dimensions would require two independent equations in x, y, z. We saw how much
of a hassel this was for something as simple as a line. I’d rather not attempt a general treatment
of the purely cartesian description of the curves in this section15 I instead offer a pair of examples
to give you a flavor:
√
Example 1.4.34. Suppose x2 + y 2 + z 2 = 4 and √ x = 2 then the solution set of this pair of
equations defines a curve in R 3 . Substutiting x = 3 into x2 + y 2 + 2 2 2
√z = 4 gives y + z = 1. The
solution set is just a unit-circle in the yz-coordinates placed at x = 3. We can parametrize it via:
√
~r(t) = h 3, cos t, sin ti.
Example 1.4.35. Suppose z = x2 − y 2 and z = 2x. The solution set is once more a curve in R 3 .
We can substitute z = 2x into z = x2 − y 2 to obtain x2 − y 2 = 2x hence x2 − 2x − y 2 = 0 and
completing the square reveals (x − 1)2 − y 2 = 1. This is the equation of a hyperbola. A natural
parametrization is given by x = 1 + cosh t and y = sinh t then since z = 2x we have z = 2 + 2 cosh t.
In total,
~r(t) = h1 + cosh t, sinh t, 2 + 2 cosh ti
We’ll explain the geometry of these calculations in the next section. Basically the idea is just
that when two surfaces intersect in R 3 we may obtain a curve.
15
which is not to say it hasn’t been done, in fact, viewing curves as solutions to equations is also a powerful
technique, but we focus our efforts in the parametric setting.
68 CHAPTER 1. ANALYTIC GEOMETRY
Example 1.4.36. A helix of radius R which wraps around the z-axis and has a slope of m is given
by:
~r(t) = hR cos(t), R sin(t), mti
Example 1.4.37. The curve parametrized by ~r(t) = ht, t2 , t3 for t ≥ 0 has scalar parametric
equations x = t, y = t2 , z = t3 and a graph
Example 1.4.38. The curve parametrized by ~r(t) = ht, t2 cos(3t), t3 sin(3t)i for t ≥ 0 has scalar
parametric equations x = t, y = t2 cos(t), z = t3 sin(t) and a graph
Example 1.4.39. The curve parametrized by ~r(t) = ht cos(3t), t2 sin(3t), t3 i for t ≥ 0 has scalar
parametric equations x = t cos(3t), y = t2 sin(3t), z = t3 and a graph
1.4. CURVES 69
We will explore the geometry of curves in the next chapter. We’ll find relatively simple calculations
which allow us to test how a curve bends within it’s plane of motion and bend off it’s plane of
motion. In other words, we’ll find a way to test if a curve lies in a plane and also how it curves
away from it’s tangential direction. These quantities are called torsion and curvature. It turns
out that these two quantities often classify a curve upto congruence in the sense of high-school
geometry. In other words, there is just one circle of radius 1 and we can rotate it and translate it
throughout R 3 . In this sense all circles in R 3 are the same. We’ve already seen in this section that
parametrization alone does not capture this concept. Why? Well there are many parametrizations
of a circle. Are those different circles? I would say not. I say there is a circle and there are many
pictures of the circle, some CW, some CCW, but so long as those pictures cover the same curve
then they are merely differing perspectives on the same object. That said, these differing pictures
are different. They are unique in their assignments of points to parameter values. The problem
of the differential geometry of curves is to extract from this infinity of parametrizations some
universal data. One seeks a few constants which invariantly characterize the curve independent of
the perspective a particular geometer has used to capture it. More generally this is the problem of
geometry. How can we classify spaces? What constants can we label a space with unambiguously?
70 CHAPTER 1. ANALYTIC GEOMETRY
1.5 surfaces
A surface in R 3 is a subset which usually looks two dimensional. There are three main viewpoints;
graphs, parametrizations or patches, and level-surfaces. As usual, the parametric view naturally
generalizes to surfaces in Rn for n > 3 with relatively little qualification. That said, we almost
without exception focus on surfaces in R 3 in this course so I focus our efforts in that direction.
This section is introductory in nature, basically this is just show and tell with a little algebra. Your
goal should be to learn the names of these surfaces and more importantly to gain a conceptual
foothold on the different ways to look at two-dimensional subsets of R 3 . Many of the diagrams in
this section were created with Maple, others perhaps Mathematica. Ask if interested.
Suppose f : dom(f ) ⊆ R 2 → R is a function then the set of all (x, y, z) such that z = f (x, y)
for some (x, y) ∈ dom(f ) is called the graph of f . Moreover, we denote graph(f ) =
{(x, y, f (x, y)) | (x, y) ∈ dom(f )}.
2 −y 2
Example 1.5.2. Let f (x, y) = xe−x . The graph looks something like:
Example 1.5.3. Let f (x, y) = −cosh(xy). The graph looks something like:
1.5. SURFACES 71
What is f ? Well, many interpretations exist. For example, f could represent the temperature at
(x, y). Or, f could represent the mass per unit area close to (x, y), this would make f a mass
density function. More generally, if you have a variable which depends by some single-valued rule
to another pair of variables then you can find a function in that application. Sometimes college
algebra students will ask, but what is a function really? With a little imagination the answer is
most anything. It could be that f is the cost for making x widgets and y gadgets. Or, perhaps
f (x, y) is the grade of a class as a function of x males and y females. Enough. Let’s get back to
the math, I’ll generally avoid cluttering these notes with these silly comments, however, you are
free to ask in office hours. Not all such discussion is without merit. Application is important, but
is not at all necessary for good mathematics.
We can add, multiple and divide functions of two variables in the same way we did for functions
of one variable. Natural domains are also implicit within formulas and points are excluded for
much the same reason as in single-variable calculus; we cannot divide by zero, take an even root
of a negative number or take a logarithm of a non-positive quantity if we wish to maintain a real
output16 . A typical example is given below:
Example 1.5.4. .
16
complex variables do give meaning to even roots and logarithms of negative numbers however, division by zero
and logarithm of zero continue to lack an arithmetical interpretation.
72 CHAPTER 1. ANALYTIC GEOMETRY
A vector valued function of a two real variables is an assignment of a vector for each
pair of real numbers in some domain D of R 2 . It’s a mapping (u, v) 7→ F~ (u, v) =
F1 (u, v), F2 (u, v), . . . , Fn (u, v) for each (u, v) ∈ D ⊂ R 2 . We say Fj : D ⊆ R 2 → R
is the j-th component function of F~ . Let S = F~ (D) then S is said to be a surface
parametrized by F~ . Equivalently, but not so usefully, we can write the scalar parametric
equations for S above as
When we define a parametrization of a surface it is important to give the formula for the patch and
the domain D of the parameters. We call D the parameter space. Usually we are interested in
the case of a surface which is embedded in R 3 so I will focus the examples in that direction. Note
however that the parametric equation for a plane actually embeds the plane in Rn for whatever n
you wish, there is nothing particular to three dimensions for the construction of the line or plane
parametrizations.
As I discussed in the plane section, a graph is given in terms of cartesian coordinates. In the case
of surfaces in R 3 you’ll often encounter the presentation z = f (x, y) for some function f . This is
an important class of examples, however, the criteria that f be a function is quite limiting.
Example 1.5.7. Let ~r(u, v) = hR cos(u) sin(v), R sin(u) sin(v), R cos(v)i for (u, v) ∈ [0, 2π] × [0, π].
In this case we have scalar equations:
It’s easy to show x2 + y 2 + z 2 = R2 and we should recognize that these are the parametric equations
which force ~r(u, v) to land a distance of R away from the origin for each choice of (u, v). Let
S = ~r(D) and recognize S is a sphere of radius R centered at the origin. If we restrict the domain
of ~r to 0 ≤ u ≤ 3π π
2 and 0 ≤ v ≤ 2 then we select just a portion of the sphere:
1.5. SURFACES 73
Notice that we could cover the whole sphere with a single patch. We cannot do that with a graph.
This is the same story we saw in the two-dimensional case in calculus II. Parametrized curves are
not limited by the vertical line test. Graphs are terribly boring in comparison to the geometrical
richness of the parametric curve. As an exotic example from 1890, Peano constructed a continuous17
path from [0, 1] which covers all of [0, 1] × [0, 1]. Think about that. Such curves are called space
filling curves. There are textbooks devoted to the study of just those curves. For example, see
Hans Sagan’s Space Filling Curves.
Example 1.5.8. Let ~r(u, v) = hR cos(u), R sin(u), vi for (u, v) ∈ [0, 2π] × R. In this case we have
scalar equations:
x = R cos(u), y = R sin(u), z = v.
It’s easy to show x2 + y 2 = R2 and z is free to range over all values. This surface is a circle at
each possible z. We should recognize that these are the parametric equations which force ~r(u, v) to
land on a cylinder of raduius R centered on the z-axis. If we restrict the domain of ~r to 0 ≤ u ≤ π
and 0 ≤ v ≤ 2 then we select a finite half-cylinder:
Example 1.5.9. Let ~r(u, v) = ha cos(u), v, b sin(u)i for (u, v) ∈ [0, 2π] × R. In this case we have
scalar equations:
x = a cos(u), y = v, z = b sin(u).
It’s easy to show x2 /a2 + z 2 /b2 = 1 and y is free to range over all values. This surface is a ellipse
at each possible y. We should recognize that these are the parametric equations which force ~r(u, v)
to land on an elliptical cylinder centered on the y-axis. If we restrict the domain of ~r to 0 ≤ u ≤ π
and 0 ≤ v ≤ 2 then we select a finite half-cylinder:
17
we will define this carefully in a future chapter
74 CHAPTER 1. ANALYTIC GEOMETRY
Example 1.5.10. Let ~r(u, v) = hR cos(u) sinh(v), R sin(u) sinh(v), R cosh(v)i for (u, v) ∈ [0, 2π] ×
R. In this case we have scalar equations:
The part of the lower branch which is graphed above is covered by the mapping Let ~r(u, v) =
hR cos(u) sinh(v), R sin(u) sinh(v), −R cosh(v)i for (u, v) ∈ [0, 2π] × [−2, 2]. The grey shape is where
the parametrization will cover if we enlarge the domain of the parameterizations.
Example 1.5.11. Let ~r(u, v) = hR cosh(u) sin(v), R sinh(u) sin(v), R cos(v)i for (u, v) ∈ R×[0, 2π].
In this case we have scalar equations:
It’s easy to show x2 − y 2 + z 2 = R2 . I’ve plotted ~r with domain restricted to dom(~r) = [−1.3, 1.3] ×
[0, π] in blue and dom(~r) = [−1.3, 1.3]×[π, 2π] in green. The grey shape is where the parametrization
will go if we enlarge the domain.
1.5. SURFACES 75
We say that S is a level surface of level k with level function F . In other words, S =
F −1 {k} is a level surface.
A level surface is a fiber of a real-valued function on R 3 .
Example 1.5.13. Let F (x, y, z) = a(x − xo ) + b(y − yo ) + c(z − zo ). Recognize that the solution
set of F (x, y, z) = 0 is the plane with base-point (xo , yo , zo ) and normal ha, b, ci.
18
to pick out a two-dimensional surface in R 4 it would take two equations in t, x, y, z, but, we really only care
about R 3 so, I’ll behave and stick with that case.
76 CHAPTER 1. ANALYTIC GEOMETRY
Example 1.5.14. some level surfaces I can plot without fancy CAS programs:
In fact, I don’t think I want to parametrize this beast. Wait, I have students, isn’t this what
homework is for?
Example 1.5.16. Let F (x, y, z) = x2 /a2 + y 2 /b2 + z 2 /c2 . The solution set of F (x, y, z) = 1 is
called an ellipsoid centered at the origin. In the special case a = b = c = R the ellipsoid is a
sphere of radius R. Here’s a special case, a = b = c = 1 the unit-sphere:
Example 1.5.17. Let F (x, y, z) = x2 + y 2 − z 2 . The solution set of F (x, y, z) = 0 is called an cone
through the origin. However, the solution set of F (x, y, z) = k for k 6= 0 forms a hyperboloid of
1.5. SURFACES 77
one-sheet for k > 0 and a hyperboloid of two-sheets for k < 0. The hyperboloids approach
the cone as the distance from the origin grows. I plot a few representative cases:
Some of the examples above fall under the general category of a quadratic surface. Suppose
For any particular nontrivial selection of constants a, b, . . . , h, i we say the solution of Q(x, y, z) = k
is a quadratic surface. For future reference let me list the proper terminology. We’d like to get
comfortable with these terms.
2. a standard elliptic paraboloid is the solution set of z/c = x2 /a2 + y 2 /b2 . If a = b then we
say the paraboloid is circular.
78 CHAPTER 1. ANALYTIC GEOMETRY
4. a standard elliptic cone is the solution set of z 2 /c2 = x2 /a2 + y 2 /b2 . If a = b then we say
the cone is circular.
5. a standard hyperboloid of one sheet is the solution set of x2 /a2 + y 2 /b2 − z 2 /c2 = 1.
6. a standard hyperboloid of two sheets is the solution set of x2 /a2 + y 2 /b2 − z 2 /c2 = −1.
1.5. SURFACES 79
If you study the formulas above you’ll notice the absense of certain terms in the general quadratic
form: terms such as dxy, exz, f yz, gx, hy are absent. Inclusion of these terms will either shift or
rotate the standard equations. However, we need linear algebra to properly construct the rotations
from the eigenvectors of the quadratic form. I leave that for Math 321 where we have more toys to
play with. You’ll have to be content with the standard examples for the majority of this course.
I’ve inserted the term standard because I don’t mean to say that every elliptic cone has the same
equation as I give. I expect you can translate the standard examples up to an interchange of
coordinates, that ought not be too hard to understand. For example, y = x2 + 2z 2 is clearly an
elliptical cone. Or y = x2 − z 2 is clearly a hyperbolic paraboloid. Or x2 + z 2 − y 2 = 1 is clearly a
hyperboloid of one sheet whereas −x2 − z 2 + y 2 = 1 is a hyperboloid of two sheets. These are the
possibilities we ought to anticipate when faced with the level set of some quadratic form. I don’t
try to memorize all of these, I use the method sketched in the next pair of examples. Basically the
idea is simply to slice the graph into planes where we find either circles, hyperbolas, lines, ellipses
or perhaps nothing at all. Then we take a few such slices and extrapolate the graph. Often the
slices x = 0, y = 0 and z = 0 are very helpful.
Example 1.5.18. .
80 CHAPTER 1. ANALYTIC GEOMETRY
Example 1.5.19. .
What about the geometry of surfaces? How can we classify surfaces? What constants can we
label a surface with unambiguously? Here’s a simple example: x2 + y 2 + z 2 = 1 and (x − 1)2 +
(y − 2)2 + (z − 3)2 = 1 define the same shape at different points in R 3 . The problem of the
differential geometry of surfaces is to find general invariants which discover this equivalence through
mathematical calculation. This is a more difficult problem and we will not treat it in this course.
It turns out this geometry begins to provide the concepts needed for Einstein’s General Relativity.
In any event, we do not currently have a course at LU which does this topic justice. I know of
1.5. SURFACES 81
at least two professors who will happily conduct an independent study on this topic once you’ve
mastered linear algebra.
Example 1.5.22. .
82 CHAPTER 1. ANALYTIC GEOMETRY
Example 1.5.23. .
Example 1.5.24. .
Example 1.5.25. .
1.6. CURVELINEAR COORDINATES 83
A coordinate system is called right-handed if the unit-vectors which point in the direction of
increasing coordinates at each point are related to each other by the right-hand-rule just like the
xyz-coordinates. We call this set of unit-vectors the frame of the coordinate system. Generally a
frame in Rn is just an assignment of n-vectors at each point in Rn . In linear-algebra langauge, a
frame is an assignment of a basis at each point of Rn . Dimensions n = 2 and n = 3 suffice for our
purposes. If y1 , y2 , y3 denote coordinates with unit-vectors u
b1 , u
b2 , u
b3 in the direction of increasing
y1 , y2 , y3 respective then we say the coordinate system is right-handed iff
b1 × u
u b2 = u
b3 , b2 × u
u b3 = u
b1 , b3 × u
u b1 = u
b2 .
x = r cos θ, y = r sin θ
Geometrically it is clear that we can label any point in R 2 either by cartesian coordinates (x, y) or
by polar coordinates (r, θ). We may view equations in cartesian or polar form.
Typically in the polar context the angle plays the role of the independent variable. In the same
way it is usually customary to write y = f (x) for a graph we try to write r = f (θ).
84 CHAPTER 1. ANALYTIC GEOMETRY
Example 1.6.2. The line y = mx + b has polar equation r sin θ = mr cos θ + b hence
b
r= .
sin θ − m cos θ
Example 1.6.3. The polar equation θ = π/4 translates to y = x for x > 0. The reason is that
π −1 y π y y
= tan ⇒ tan = ⇒ 1= ⇒ y=x
4 x 4 x x
and the ray with θ = π/4 is found in quadrant I where x > 0.
Let us denote unit vectors in the direction of increasing r, θ by rb, θb respective. You can derive by
geometry alone that
rb = cos(θ) x
b + sin(θ) yb
(1.5)
θ = − sin(θ) x
b b + cos(θ) yb.
We call {b r, θ}
b the frame of polar coordinates. Notice that these are perpendicular at each point;
rb • θ = 0.
b
Example 1.6.4. If we want to assign a vector to each point on the unit circle such that the vector
is tangent and pointing in the counter-clockwise (CCW) direction then a natural choice is θ.
b
Example 1.6.5. If we want to assign a vector to each point on the unit circle such that the vector
is pointing radially out from the center then a natural choice is rb.
Suppose you have a perfectly flat floor and you pour paint slowly in a perfect even stream then in
principle you’d expect it would spread out on the floor in the rb direction if we take the spill spot
as the origin and the floor as the xy-plane.
2 2 2 −1 y
r =x +y , θ = π + tan .
x
Geometrically it is clear that we can label any point in R 3 either by cartesian coordinates (x, y, z)
or by cylindrical coordinates (r, θ, z).
Example 1.6.6. In cylindrical coordinates the equation r = 1 is a cylinder since the z-variable is
free. If we denote the unit-circle by S1 = {(x, y) | x2 + y 2 = 1} then the solution set of r = 1 has
the form S1 × R. At each z we get a copy of the circle S1 .
Example 1.6.7. The equation θ = π/4 is a half-plane which has equation y = x subject to the
condition x > 0.
Example 1.6.8. Suppose we have a line of electric charge smeared along the z-axis with charge
density λ. One can easily derive from Gauss’ Law that the electric field has the form:
~ = λ
E rb.
2πo r
Example 1.6.9. If we have a uniform current I zb flowing along the z-axis then the magnetic field
can be derived from Ampere’s Law and has the simple form:
~ = µo I θb
B
2πr
Trust me when I tell you that the formulas in terms of cartesian coordinates are not nearly as clean.
If we fix our attention to a particular point the cylindrical frame has the same structure as the
cartesian fram { x
b, yb, zb}. In particular, we can show that
rb • rb = 1, θb • θb = 1, zb • zb = 1
θb • rb = 0, θb • zb = 0, zb • rb = 0.
We can also calculate either algebraically or geometrically that:
rb × θb = zb, θb × zb = rb, zb × rb = θb
Therefore, the cylindrical coordinate system (r, θ, z) is a right-handed coordinate system since
it provides a right-handed basis of unit-vectors at each point. We can summarize these relations
compactly with the notation ub1 = rb, u
b2 = θ,
b ub3 = zb whence:
3
X
u
bi • u
bj = δij , bi × u
u bj = ijk u
bk
k=1
this is the same pattern we saw for the cartesian unit vectors.
1.6. CURVELINEAR COORDINATES 87
x = ρ cos(θ) sin(φ)
y = ρ sin(θ) sin(φ) (1.7)
z = ρ cos(φ)
ρ2 = x2 +py2 + z2
tan(φ) = x2 + y 2 /z (1.8)
tan(θ) = y/x.
It is clear that any point in R 3 is labeled both by cartesian coordinates (x, y, z) or spherical
coordinates (ρ, φ, θ).
Also, it is important to distinguish between the geometry of the polar angle θ and the azmuthial
angle φ
19
I’ll use notation which is consistent with Stewart, but beware there is a better notation used in physics and
engineering where the meaning of φ and θ are switched and the spherical radius ρ is instead denoted by r
88 CHAPTER 1. ANALYTIC GEOMETRY
p
Example 1.6.10. The equation x2 + y 2 + z 2 = R is written as ρ = R in spherical coordinates.
Example 1.6.11. The plane a(x − 1) + b(y − 2) + c(z − 3) = 0 has a much uglier form in spherical
coordinates. Its: a(ρ cos(θ) sin(φ) − 1) + b(ρ sin(θ) sin(φ) − 2) + c(ρ cos(φ) − 3) = 0 hence
a + 2b + 3c
ρ= .
a cos(θ) sin(φ) + b sin(θ) sin(φ) + c cos(φ)
x2 + y 2 = ρ2 cos2 (θ) sin2 (φ) + ρ2 cos2 (θ) sin2 (φ) = ρ2 sin2 (φ)
You might notice that the formula above is easily derived geometrically. If you picture a cylinder
and draw a rectangle as shown below it is clear that sin(φ) = Rρ .
It is important to be proficient in both visualization and calculation. They work together to solve
problems in this course, if you get stuck in one direction sometimes the other will help you get free.
Let us denote unit vectors in the direction of increasing ρ, φ, θ by ρb, φ,
b θb respectively. You can
derive by geometry alone that
ρb = sin(φ) cos(θ) x
b + sin(φ) sin(θ) yb + cos(φ) zb
φb = − cos(φ) cos(θ) x
b − cos(φ) sin(θ) yb + sin(φ) zb (1.9)
θ = − sin(θ) x
b b + cos(θ) yb.
We call {bρ, φ,
b θ}b the frame of spherical coordinates. At each point these unit-vectors point in
particular direction.
1.6. CURVELINEAR COORDINATES 89
Example 1.6.13. Suppose we a point charge q is placed at the origin then by Gauss’ Law we can
derive
~ = q
E ρb.
4πo ρ2
This formula makes manifest the spherical direction of the electric field, the absence of the angular
unit-vectors says the field has no angular dependence and hence its values depend only on the
spherical radius ρ. This is called the Coulomb field or monopole field. Almost the same math
applies to gravity. If M is placed at the origin then
GmM
F~ = (−b
ρ).
ρ2
gives the gravitational force F~ of M on m at distance ρ from the origin. The direction of the
gravitational field is −b
ρ which simply says the field points radially inward.
The spherical frame gives us a basis of vectors to build vectors at each point in R 3 . More than
that, the spherical frame is an orthonormal frame since at any particular point the frame provides
an orthonormal set of vectors. In particular, we can show that
ρb • ρb = 1, φb • φb = 1, θb • θb = 1
φb • ρb = 0, θb • ρb = 0, φb • θb = 0.
We can also calculate either algebraically or geometrically that:
θb × ρb = φ,
b ρb × φb = θ,
b φb × θb = ρb
20
b = h1, 0, 0i, yb = h0, 1, 0i and
How do I know the cartesian frame is unchanging? It’s not complicated really; x
zb = h0, 0, 1i.
90 CHAPTER 1. ANALYTIC GEOMETRY
Therefore, the spherical coordinate system (ρ, φ, θ) is a right-handed coordinate system since it
provides a right-handed basis of unit-vectors at each point. We can summarize these relations
compactly with the notation ub1 = ρb, u
b2 = φ,
b ub3 = θb whence:
3
X
u
bi • u
bj = δij , bi × u
u bj = ijk u
bk
k=1
this is the same pattern we saw for the cartesian unit vectors.
We will return to the polar, cylindrical and spherical coordinate systems as the course progresses.
Even now we could consider a multitude of problems based on the combination of the material cov-
ered thus-far and it’s intersection with curvelinear coordinates. There are other curved coordinate
systems beyond these standard three, but I leave those to your imagination for the time being. I do
discuss a more general concept of coordinate system in the advanced calculus course. In manifold
theory the concept of a coordinate system is refined in considerable depth. We have no need of
such abstraction here so I’ll behave21 .
21
I’d guess most calculus text editors would say this whole paragraph is misbehaving, but I have no editor so ha.