Ch 4 Notes Packet 2022.Doc
Ch 4 Notes Packet 2022.Doc
We can think of a vector space in general, as a collection of objects that behave as vectors do in Rⁿ. The objects of such a set are
called vectors.
A vector space is a nonempty set V of objects, called vectors, on which are defined two operations, called addition and multiplication
by scalars (real numbers), subject to the ten axioms below. The axioms must hold for all u, v and w in V and for all scalars c and d.
1. u + v is in V.
2. u + v = v + u
3. (u + v) + w = u + (v + w)
4. There is a vector (called the zero vector) 0 in V such that u + 0 = u.
5. For each u in V, there is vector -u in V satisfying u + (-u) = 0.
6. cu is in V.
7. c(u + v) = cu+cv.
8. (c + d)u = cu+du.
9. (cd)u = c(du).
10. 1u = u.
Rn—the set of all column (or row vectors with n components where the components are real numbers
Vector spaces where the scalars are complex numbers are called complex vector spaces. If the scalars are real, they are called real
vector spaces.
The simplest of all vector spaces is the zero vector space.
Ex 2: Let n ≥ 0 be an integer and let Pn = the set of all polynomials of degree at most n ≥ 0.
Members of Pn have the form p(t)=a₀ + a₁t + a₂t² + ⋯ + antn where a₀,a₁,…,an are real numbers and t is a real variable. The set Pn is
a vector space. To prove it is a vector space, we must prove that all 10 axioms hold. We will just verify 3 out of the 10 axioms here.
Let p(t)=a₀ + a₁t + a₂t² + ⋯ + antn and q(t)=b₀ + b₁t + b₂t² + ⋯ + bntn . Let c be a scalar.
Axiom 1 : The polynomial p + q is defined as follows: (p + q)(t)=p(t)+q(t). Therefore, (p + q)(t)=p(t)+q(t) = (______) + (______)t
(p + 0)(t) = p(t) + 0 = (a₀+ 0)+(a₁+ 0)t + ⋯ + (an + 0)tⁿ = a₀ + a₁t + ⋯ + an tⁿ = p(t) and so p + 0 = p
Subspaces: Vector spaces may be formed from subsets of other vectors spaces. These are called subspaces.
A subspace of a vector space V is a subset H of V that has three properties:
If the subset H satisfies these three properties, then H itself is a vector space.
The smallest subspace is the zero vector.
The empty set is not a subspace.
Ex 4: Is a subspace of R2?
b. To show that Span{ v₁, …, vp } is closed under vector addition, we choose two arbitrary vectors in Span{ v₁, …, vp }:
u = a₁v₁+ a₂v₂ + ⋯ + anvn and v = b₁v₁+ b₂v₂ + ⋯ + bnvn Then
u + v = (__________________________) + (__________________________) =
c. To show that Span{ v₁, …, vp } closed under scalar multiplication, choose an arbitrary number c and an arbitrary vector in Span{
v₁, …, vp }, v = b₁v₁+ b₂v₂ + ⋯ + bnvn
Recap
Ex 5: Is V={(a + 2b, 2a - 3b): a and b are real} a subspace of R²? Why or why not?
Theorem 2
The null space of an m x n matrix A is a subspace of Rn. Equivalently, the set of all solutions to a system Ax = 0
of m homogeneous linear equations in n unknowns is a subspace of Rn.
Proof: Nul A is a subset of Rn since A has n columns. Must verify properties a, b, and c of the definition of a subspace.
Observations:
1. Spanning set of Nul A, found using the method in the last example, is automatically linear independent:
2. If , then the number of vectors in the spanning set for Nul A equals the number of free
variables in Ax = 0.
The column space of an m x n matrix A (Col A) is the set of all linear combinations of the columns of A.
Recall that if Ax = b, then b is a linear combination of A. Therefore,Col A = {b: b = Ax for some x in Rn}
The column space of an m x n matrix A is all of Rm if and only if the equation Ax = b has a solution for each b in Rm.
Why?
Ex 3: Is w in Nul A?
Ex 4: Is w in Col A?
Ex 5: Let
(a) The column space of A is a subspace of Rk where k = _______________.
(c) Find a nonzero vector in Col A. (There are infinitely many possibilities.)
(d) Find a nonzero vector in Nul A. Solve Ax = 0 and pick one solution.
x1 = -2x2, x2 is free, x3 = 0
Contrast Between Nul A and Col A for an m x n matrix A (see page 204)
Nul A Col A
1. Subspace of Rn 1. Subspace of Rm
2. Implicit 2. Explicit
3. Takes time to find vectors in Nul A 3. Easy to find vectors in Col A
4. No obvious relationship between Nul A 4. Obvious relationship between Col A and entries in A
and the entries in A
5. Av = 0, v in Nul A 5. Ax = v consistent, v in Col A
6. Easy to tell if v is in Nul A, compute Av 6. May take time to tell if v is in Col A, Do [A v]
7. Nul A = {0} iff Ax = 0 has only trivial solution 7. Col A = Rm iff Ax = b has a solution for every b in Rm
8. Nul A = {0} iff linear transformation 8. Col A = Rm iff linear transformation
is one-to-one maps Rn onto Rm
Review:
A subspace of a vector space V is a subset H of V that has three properties:
a. The zero vector of V is in H.
b. For each u and v in H, u + v is in H.
(closed under addition)
c. For each u in H and each scalar c, cu is in H.
(closed under scalar multiplication)
If the subset H satisfies these three properties, then H itself is a vector space.
Ex 6: Determine whether each of the following sets is a vector space or provide a counterexample.
(a)
(b)
(c)
A linear transformation T from a vector space V into a vector space W is a rule that assigns to each x in V a unique vector
T(x) in W, such that
i. T(u + v) = T(u) + T(v) for all u,v in V
ii. T(cu) = cT(u) for all u in V and all scalars c.
The kernel (or null space) of T is the set of all vectors u in V such that T(u) = 0.
The range of T is the set of all vectors in W of the form T(u) where u is in V.
Ex 8: mapping to .
Is (0, 0, 2) in ker L?
Ex 9: Let A = and w =
Is w in Nul A?
Is w in Col A?
A set of vectors { v₁, …, vp } in a vector space V is said to be linearly independent if the vector equation
c1 v₁+ c2 v2 + …+ cp vp = 0 has only the trivial solution c₁=0,…,cp =0.
The set { v₁, …, vp } is said to be linearly dependent if there exists weights c₁=0,…,cp, not all 0, such that
c1 v₁+ c2 v2 + …+ cp vp = 0
The following results from Section 1.7 are still true for more general vectors spaces.
A set of two vectors is linearly dependent if and only if one is a multiple of the other.
Theorem 4: An indexed set { v₁, …, vp } of two or more vectors, with vj ≠ 0, is linearly dependent if and only if some vector vj (j >1)
is a linear combination of the preceding vectors v₁, …, vj-1
Ex 3: Let {p₁, p₂, p₃} be a set of vectors in P₂ where p₁(t) = t, p₂(t) = t², and p₃(t) = 4t + 2t². Is this a linearly dependent set?
A Basis Set Let H be the plane illustrated below. Which of the following are valid descriptions of H?
(a) H=Span{v₁,v₂}
(b) H=Span{v₁,v₃}
(c) H=Span{v₂,v₃}
(d) H=Span{v₁,v₂,v₃}
A basis set is an "efficient" spanning set containing no unnecessary vectors. In this case, we would consider the linearly independent
sets {v₁,v₂} and {v₁,v₃} to both be examples of basis sets or bases (plural for basis) for H.
Definition: Let H be a subspace of a vector space V. An indexed set of vectors β={b₁,…,bp} in V is a basis for H if
Ex 4: Let e₁= , e2= , e3= . Show that {e₁,e₂,e₃} is a basis for R³. The set {e₁,e₂,e₃} is called a standard basis for R³.
(Hint: Use the IMT)
Ex 7: Explain why each of the following sets is not a basis for R³.
(a) (b)
In this example, the set of three vectors {v1, v2, v3} is a spanning set for Nul A. In the last section we observed that this set is linearly
independent. Therefore {v1, v2, v3} is a basis for Nul A. The technique used here always provides a linearly independent set.
The Spanning Set Theorem: Informally, this theorem says a basis can be constructed from a spanning set of vectors by discarding
vectors which are linear combinations of preceding vectors in the indexed set.
Theorem 5 (The Spanning Set Theorem): Let S={ v₁, …, vp } be a set in V and let H= Span{ v₁, …, vp }
a. If one of the vectors in S, say vk, is a linear combination of the remaining vectors in S, then the set formed from S by removing vk
still spans H.
b. If H ≠ {0}, some subset of S is a basis for H.
Review:
1. To find a basis for Nul A, use elementary row operations to transform [A 0] to an equivalent reduced row echelon form [B 0]. Use
the reduced row echelon form to find parametric form of the general solution to Ax=0. The vectors found in this parametric form of
the general solution form a basis for Nul A.
2. A basis for Col A is formed from the pivot columns of A. Warning: Use the pivot columns of A, not the pivot columns of B, where
B is in reduced echelon form and is row equivalent to A.
In general, people are more comfortable working with the vector space Rⁿ and its subspaces than with other types of vectors spaces
and subspaces. The goal here is to impose coordinate systems on vector spaces, even if they are not in Rⁿ.
Theorem 7 The Unique Representation Theorem: Let β={b₁,…, bn} be a basis for a vector space V. Then for each x in V, there
exists a unique set of scalars c₁,…,cn such that x = c₁b₁ + ⋯ + cnbn
Definition: Suppose β={b₁,…, bn} is a basis for a vector space V and x is in V. The coordinates of x relative to the basis β (or the β-
coordinates of x) are the weights c₁,…,cn such that x = c₁b₁ + ⋯ + cnbn
In this case, the vector in Rⁿ [x]β = is called the coordinate vector of x (relative to β), or the β- coordinate vector of x
Ex 1: Let β={b₁,b₂} where b₁= and b₂= and let E={e₁,e₂} where e₁=
and e₂= . Express [x]β = in terms of the E={e₁,e₂}; i.e., find [x]E.
Geometrically, we are changing from using basis vectors <3,1> and <0,1> to standard basis vectors <1, 0> and <0, 1>.
Standard graph paper β- graph paper
For a basis β = {b₁,. . .,bn} let Pβ = [b₁,. . .,bn] and [x]β = Then x = Pβ [x]β
We call Pβ the change-of-coordinates matrix from β to the standard basis in Rⁿ. Then [x]β = Pβ-1 x
and therefore Pβ-1 is a change-of-coordinates matrix from the standard basis in Rⁿ to the basis β.
Ex 2: Let β={b₁,b₂} where b₁= , b₂= and x = . Find the change-of-coordinates matrix Pβ from β to the standard basis
in R² and change-of-coordinates matrix Pβ-1 from the standard basis in R² to β.
Ex 3: Find the vector x determined by the given coordinate vector [x]β and the given basis β.
β= , [x]β =
Ex 4: Find the coordinate vector [x]β of x relative to the given basis β = {b₁,. . .,bn} b1 = , b2 = ,x =
Coordinate mappings allow us to introduce coordinate systems for unfamiliar vector spaces.
Standard basis for P₂:{p₁, p₂,p₃}= {1, t, t²}
Informally, we say that vector space V is isomorphic to W if every vector space calculation in V is accurately reproduced in W, and
vice versa.
Assume β is a basis set for vector space V. Exercise 25 (page 223) shows that a set {u₁,u₂,…,up} in V is linearly independent if and
only if {[u₁]β,[u₂]β,…,[up] β} is linearly independent in Rⁿ.
Ex 5: Use coordinate vectors to determine if {p₁, p₂, p₃} is a linearly independent set, where p₁= 1 – t , p₂= 2 – t + t², and p₃= 2t +
3t².
Ex 6: Determine whether the set of polynomials forms a basis for P3. {5 – 3t + 4t2 + 2t3, 9 + t + 8t2 – 6t3,
6 – 2t + 5t2, t3}
Coordinate vectors also allow us to associate vector spaces with subspaces of other vectors spaces.
Ex 7: Let β={b₁,b₂} where b₁= ,b₂= and let H=span{b₁,b₂}. Find [x]β, if x=
Theorem 9: If a vector space V has a basis β = {b1, . . ., bn}, then any set in V containing more than n vectors must be linearly
dependent.
Proof: Suppose is a set of vectors in V where p > n. Then the coordinate vectors are in . Since p >
Theorem 10: If a vector space V has a basis of n vectors, then every basis of V must consist of n vectors.
Proof: Suppose is a basis for V consisting of exactly n vectors. Now suppose is any other basis for V. By definition of a
basis, we know that and are both linearly independent sets.
By Theorem 9, if has more vectors than , then _______ is a linearly dependent set (which cannot be the case).
Again by Theorem 9, if has more vectors than , then _________ is a linearly dependent set (which cannot be the case).
Definition – If V is spanned by a finite set, then V is said to be finite-dimensional, and the dimension of V, written as dim V, is the
number of vectors in a basis for V. The dimension of the zero vector space {0} is defined to be 0. If V is not spanned by a finite set,
then V is said to be infinite-dimensional.
The standard basis for is where are the columns of . So, for example, dim = 3.
Dimensions of subspaces of
2-dimensional subspaces. Span {u, v} where u and v are in R3 and are not multiples of each other. These subspaces are
3-dimensional subspaces. Span {u, v, w} where u, v, w are linearly independent vectors in . This subspace is itself because
the columns of A = [u v w] span according to the IMT.
Theorem 11: Let H be a subspace of a finite-dimensional vector space V. Any linearly independent set in H can be expanded, if
necessary, to a basis for H. Also, H is finite-dimensional and .
Let . Then H is a subspace of and dim H < dim . We could expand the spanning set to
Theorem 12--The Basis Theorem: Let V be a p-dimensional vector space, . Any linearly independent set of exactly p vectors in
V is automatically a basis for V. Any set of exactly p vectors that spans V is automatically a basis for V.
Dimensions of Col A and Nul A: Recall our techniques to find basis sets for column spaces and null spaces.
4.6 Rank
The set of all linear combinations of the row vectors of a matrix A is called the row space of A and is denoted by Row A.
For example, let and r1 = (-1, 2, 3, 6), r2 = (2, -5, -6, -12), r3 = (1, -3, -3, -6)
Row A = Span{ r1 , r2 , r3 }(a subspace of R4)
While it is natural to express row vectors horizontally, they can also be written as column vectors if it is more convenient.
When we use row operations to reduce matrix A to matrix B, we are taking linear combinations of the rows of A to come up with B.
We could reverse this process and use row operations on B to get back to A. Because of this, the row space of A equals the row space
of B.
Theorem 13: If two matrices A and B are row equivalent, then their row spaces are the same. If B is in echelon form, the nonzero rows
of B form a basis for the row space of A as well as B.
Ex 1: The matrices are row equivalent. Find a basis for row space, column space
and null space of A. Also state the dimension of each.
# of pivot # of nonpivot # of
columns columns columns
of A of A of A
Theorem 14 The Rank Theorem: The dimensions of the column space and the row space of an m x n matrix A are equal. This
common dimension, the rank of A, also equals the number of pivot positions in A and satisfies the equation rank A + dim Nul A = n
Ex 2: Suppose that a 5 x 8 matrix A has rank 5. Find dim Nul A, dim Row A and rank . Is Col A = ?
Ex 4: A scientist solves a homogeneous system of 50 equations in 54 variables and finds that exactly 4 of the unknowns are free
variables. Can the scientist be certain that any associated nonhomogeneous system (with the same coefficients) has a solution?
Ex 5: If a 6x3 matrix A has rank 3, find dim Nul A, dim Row A and rank AT.
Ex 7: If the null space of a 7 x 6 matrix A is 5 dimensional, what is the dimension of the column space of A?
Ex 8: If the Null Space of a 5 x 6 matrix A is 4-dimensional, what is the dimension of the row space of A?
Theorem 15: Let B = {b1, . . . , bn} and C = {c1, . . . , cn} be bases of a vector space V. Then there is a uniquen x n matrix such
The columns of are the C-coordinate vectors of the vectors in the basis B. That is, = [ [b1]C [b2]C ... [bn]C ]
Ex 1: Consider two bases B = {b1, b2} and C = {c1, c2} for a vector space V
such that b1 = -c1 + 4c2 and b2 = 5c1 - 3c2
The columns of are linearly independent because they are the coordinate vectors of the linearly independent set B.
Since is square, it must be invertible, by the IMT.
To change coordinates between two nonstandard bases in Rn, we need Theorem 15. The theorem shows that to solve the
change-of-basis problem, we need the coordinate vectors of the old basis relative to the new basis.
In Ex 2 and 3, let B = {b1, b2} and C = {c1, c2} be bases for R2. Find the change-of-coordinates matrix from B to C and the
change-of-coordinates matrix from C to B. (Hint: For , row reduce [c1 c2 b1 b2] and
Note: or
Ex 4: In P2, find the change-of-coordinates matrix from the basis to the standard basis. Then write
as a linear combination of the polynomials in B.
If the probability of a given state can be predicted by just knowing the state of the system at the preceding observation, then the
process of change is called a Markov chain or Markov process.
Definition: If a Markov chain has k possible states, which we label as 1, 2, . . . k, then the probability that the system is in state I at
any observation after it was in state j at the preceding observation is denoted by pij and is called the transition probability from state j
to state i. The matrix P = [pij] is called the transition matrix of the Markov chain.
For example, in a three state Markov chain, the transition matrix has the form
Preceding State
1 2 3
There is a second important matrix in the use of Markov chains, the state vector.
Definition: The state vector for an observation of a Markov chain with k states is a column vector x whose ith component xi is the
probability that the system is in the ith state at that time.
Consider the following example: Rent-a-Lemon has three locations from which to rent a car for one day: airport, downtown and the
valley
Daily Migration:
Rented From
Airport Downtown Valley Returned to
.95 .02 .05 Airport
.03 .90 .05 Downtown
.02 .08 .90 Valley
M= (migration matrix)
Note that the probabilities in each column total one. A matrix with this property is called a stochastic matrix.
If we wish to find the distribution of cars after one day, we can find
In general,
since x = Mx
Ex 1: By reviewing its donation records, the alumni office of a college finds that 80% of its alumni who contribute to the annual fund
one year will also contribute the next year, and 30% of those who do not contribute one year will contribute the next. Construct a
transition matrix and initial state matrix illustrating this situation. Find the contributions for each of the first five years. Continue until
you find the steady state vector.
There is process that can be used to find the steady state vector so that it is not necessary to use trial and error.
Mx – Ix = 0
Factoring, we get
(M – I)x = 0
Ex 2: Find the steady state vectors for each of the following transition matrices
a.
b.
Ex 3: Suppose that 3% of the population of the U.S. lives in the State of Washington. Suppose the migration of the population into
and out of Washington State will be constant for many years according to the following migration probabilities. What percentage of
the total U.S. population will eventually live in Washington?
From
WA Rest of U.S. To:
0.9 0.01 WA
0.10 0.99 Rest of U. S.