Home
Syllabus
Lecture Topics
Homework
Test Solutions
Policies
|
|
Math 3130: Introduction to
Linear Algebra,
Spring 2013
|
|
Lecture Topics
|
|
|
Date
|
What we discussed/How we spent our time
|
Jan 14
|
Syllabus. Text.
Section 1.1:
Systems of linear equations. Augmented matrix of a linear system.
Elementary row operations.
|
Jan 16
|
Section 1.2:
Row reduction. (Reduced) row echelon form.
Pivots, pivot positions and pivot columns.
Free and pivot variables. Solution sets.
|
Jan 18
|
We reviewed terminology from Sections 1.1 and 1.2.
(row reduction/Gaussian elimination, row echelon form,
reduced row echelon form, augmented matrix, coefficient matrix,
consistent vs. inconsistent system, existence and uniqueness of solutions,
overdetermined and underdetermined systems).
Then we began discussing matrix arithmetic
(addition, zero matrix additive inverse,
multiplication, identity matrix, multiplicative inverses)
to prepare to discuss the matrix point of view
for solving linear systems,
described in Sections 1.3 and 1.4.
|
Jan 23
|
We discussed matrix arithmetic, including the
fact that the collection of nxn real matrices
forms a ring (which is noncommutative when
n is greater than 1). We explained how linear
systems are expressible as matrix equations of the form
Ax = b.
Quiz 1.
Read Sections 1.3 and 1.4, and skim Sections 2.1 and 2.2.
|
Jan 25
|
We explained that row reduction
corresponds to left multiplication by certain
invertible matrices.
We defined elementary matrices
Eii(r), Eji(r), Pij,
and general permutation matrices.
We explained how to find the inverse of each elementary matrix,
and how to find the inverse of a product of invertible matrices.
|
Jan 28
|
We defined real vector spaces, and
gave geometric representations for vectors
and for vector arithmetic (adding and scaling) in
ℝn.
We defined "linear combination" and "span".
We recast the problem of solving a linear system
as solving a vector equation and also as the problem
of finding the weights needed to express
a given vector as linear combination of others.
Quiz 2.
|
Jan 30
|
We discussed the connection between linear systems,
matrix equations and vector equations.
We gave informal definitions of "vector space"
and "subspace".
|
Feb 1
|
Read Section 1.5.
We showed that the solution set of a homogeneous
system with n unknowns is a subspace of ℝn,
called the nullspace of A,
and that the set of all vectors b such that
Ax=b is consistent is a subspace
of ℝm, called the column space
of A. We showed that the general solution to
Ax=b may be expressed in terms of
the general solution to the associated homogeneous
system, Ax=0, along with a particular
solution p to Ax=b:
x = p + (t1*s1+
…+tk*sk).
The purpose of this representation is to show that the
general solution of a linear system may be viewed
as a translate of a subspace.
|
Feb 4
|
Read Section 1.7.
We defined linear (in)dependence. The set of columns of a matrix A
are linearly independent iff Ax=0 has only the
trivial solution. A subset of a linearly independent set
is linearly independent. The empty set of vectors is
linearly independent. A 1-element set of vectors, {v},
is linearly independent iff v is nonzero.
A 2-element set of vectors is linearly independent
iff neither vector is a multiple of the other. An n-element set of
vectors is linearly independent iff no one is a linear combination
of the others. ℝn does not contain a
linearly independent set of size greater than n, but does contain
a linearly independent set of size m for any m satisfying
0≤m≤n.
Quiz 3.
|
Feb 6
|
Read Sections 1.8-1.9.
I defined "linear transformation". (Alternative terminology:
"vector space homomorphism".) I gave
examples and nonexamples.
I showed that every linear transformation
T:ℝn→
ℝm has the form
T(x) = Ax for some uniquely determined
(m×n) matrix A, which we showed how to calculate.
We examined some examples. One nice example
(which I think is worth memorizing) is that the
counterclockwise rotation of the plane through an
angle θ is a linear transformation from
ℝ2 to
ℝ2 whose matrix is:
.
I explained why a linear transformation
T(x) = Ax is 1-1 iff Ax = 0
has only the trivial solution.
As class ended, I posed the problem
of identifying what property of the matrix A
corresponds to the property that the transformation T is onto.
|
Feb 8
|
We worked on
practice problems.
(We discussed the answers to the first four problems in class.)
|
Feb 11
|
Read Sections 1.6, 1.10
Introduction to Mathematica: I discussed
how to get
Mathematica and how to start creating a notebook.
Here is a good video tutorial series.
I worked through
a simple example using Mathematica
to balance a chemical equation.
I discussed using linear algebra to solve network flow problems,
and to study predator-prey dynamics.
Quiz 4.
|
Feb 13
|
Reread Sections 2.1-2.2
We discussed the solutions to Quiz 4, especially
what it means to "define" a term or phrase.
Then we discussed the transpose of a matrix, and the
laws of arithmetic related to transpose.
The most important (or unexpected) law is that
(AB)t=BtAt.
(Note: (AB)t≠AtBt, usually. In fact,
the dimensions of the matrices usually don't match up properly
for this to even make sense.) Finally we discussed left, right
and 2-sided inverses of a matrix. We showed that if A
has a left inverse, then Ax=0 has only the
trivial solution. More about this on Friday …
|
Feb 15
|
We practiced writing proofs using
this handout.
Our practice was prefaced by a brief discussion of (i)
how to prove the equivalence of a sequence of statements, and (ii) proof by
contradiction.
|
Feb 18
|
Read Section 2.3
We continued our discussion of inverses, focusing on the geometric
interpretation of "A is left invertible" and "A is right invertible".
We added to earlier characterizations of left/right invertibility
by proving that
A is left invertible iff T(x)=Ax
preserves independent sets, and
A is right invertible iff T(x)=Ax
preserves spanning sets. We observed that a left invertible
matrix is right invertible iff it is square.
We showed that if A is square, then Gaussian elimination
applied to [A|I] yields [I|A-1].
Quiz 5.
|
Feb 20
|
Read Sections 2.4-2.5
We first discussed the arithmetic of partitioned matrices.
(Everything works as expected, with the obvious restrictions.)
We focused briefly on block diagonal matrices and block
upper triangular matrices, noting that a matrix of either type
is invertible iff its diagonal blocks are invertible.
Second, we discussed the fact that Gaussian elimination
applied to a matrix A produces a factorization A=LR
where L is a square invertible matrix and R is in echelon form.
If Gaussian elimination can be performed without row switches
or row scaling, then L is lower triangular with 1's on the diagonal.
If A is square, then R is upper triangular (in which case we call
it U instead of R). Thus, if Gaussian
elimination applied to a square matrix A without
row switches or row scalings succeeds, then it
produces a factorization A=LU with
(i) L lower triangular with 1's on the diagonal, and
(ii) U upper triangular.
We observed that if A=LU, then the single system
Ax=b is equivalent to a pair of simpler
systems
Ly=b and Ux=y.
We mentioned that if A is n×n, then it takes
O(n3)
arithmetic operations to solve
Ax=b for a given b
via Gaussian elimination. It takes
O(n3)
arithmetic operations to factor A as LU.
But it takes only
O(n2)
arithmetic operations to solve each of
Ly=b and Ux=y.
Thus, if you want to solve many systems of the form
Ax=b for the same A but different
b's, then it can be more efficient to
factor A first.
|
Feb 22
|
We continued our discussion of LU factorizations,
and saw by example how to find one. We observed that
the inverse of a (unit) lower triangular matrix is
a (unit) lower
triangular matrix,
and the product of two (unit) lower triangular matrices
is a (unit) lower triangular matrix.
We introduced the Leontief economic model,
x=Cx+d,
which depicts relationships between different
sectors of an economy.
|
Feb 25
|
We discussed the relationship between fixed point
problems (x=F(x) for some function F)
and problems about finding roots
(0=G(x) for some function G).
In particular, the equation
x=Cx+d from the Leontief model
can be treated as an ordinary linear system
(I-C)x=d , in which case the solution is
x=(I-C)-1d, or may be treated
as a fixed point problem
x=F(x) for F(x)=
Cx+d,
in which case the solution obtained by iteration
is (I+C+C2 +…)d. These solutions are equal,
since
(I-C)-1=(I+C+C2 +…) when C
is "sufficiently small". Here the fact that the entries of C are
nonnegative and the column sums of C are less than 1
make C "sufficiently small".
Quiz 6.
|
Feb 27
|
Midterm review.
(George made a
solutions page.)
|
Mar 1
|
Midterm.
|
Mar 4
|
Read Section 2.7.
We discussed affine transformations, and how to represent them
in homogeneous coordinates.
|
Mar 6
|
Read Sections 2.7-2.9.
We discussed
subspaces of ℝn, Col(A), Row(A), Nul(A),
ordered bases for a space, standard basis for ℝn,
dimension,
rank and nullity of a matrix.
We gave algorithms for finding bases for Col(A), Row(A), Nul(A).
|
Mar 8
|
Read Sections 3.1-3.2.
We proved the rank + nullity theorem and that
dimension is well defined.
We worked on some
practice problems.
|
Mar 11
|
We discussed the definition of the determinant,
its use in caculating signed volume, and the
Laplace expansion.
Quiz 7.
|
Mar 13
|
We discussed properties of the determinant
including: the determinant is defined only for square matrices,
the determinant of a matrix is nonzero iff its columns are independent,
det(A)=det(AT), and A*adj(A) = det(A)*I (hence
A-1=(1/det(A))*adj(A) if A is invertible).
|
Mar 15
|
Read Section 3.3.
We discussed properties of the determinant
including: the determinant of (block) triangular matrices,
determinant of a product is the product of determinants,
the determinant can be computed via Gaussian elimination,
if T(x)=Ax, then the determinant of A
measures the "volume expansion" associated with T,
and that the determinant is the unique alternating
multilinear function d of n variables defined
on ℝn for which
d(e1,…,en)=1.
We briefly discussed Cramer's Rule.
|
Mar 18
|
We worked out a 3×3 example illustrating Cramer's Rule,
then discussed the permutation expension of the determinant.
Quiz 8.
|
Mar 20
|
Read Sections 4.1-4.4.
We discussed the word "abstract", then
defined (real) vector spaces. We discussed the
new examples Mm×n(ℝ),
Pn(t), and Ck([0,1]).
We computed that
Mm×n(ℝ)
has dimension mn,
Pn(t) has dimension n+1, and
C0([0,1]) is infinite dimensional.
|
Mar 22
|
Read Sections 4.5-4.6.
We discussed coordinates relative to a basis, and found coordinates
for certain vectors in
M2×2(ℝ)
and
P3(t). We defined the (very important!) terms
isomorphism
and isomorphic for real vector spaces.
We proved the (very important!) theorem
that any finitely generated real vector
space is isomorphic to ℝn for some n.
|
Apr 1
|
Read Section 4.7.
We reviewed the proof
that any finitely generated real vector
space is isomorphic to ℝn for some n.
We then explained why, if V and W are finitely generated
vector spaces with ordered bases B and C respectively
and T:V→W is a linear tranformation, then there
is a unique matrix
C[T]B
such that
C[T]B
[v]B=[T(v)]C.
We calculated this matrix for some examples, namely the matrix for
the transpose operation from
M2×2(ℝ) to itself and the matrix
for the derivative operation from
Pn(t) to
Pn-1(t).
We mentioned that if I:V→V
is the identity transformation,
then C[I]B is the change of basis matrix from
basis B to basis C.
|
Apr 3
|
We discussed change of basis a bit more, then worked
on these
practice problems.
|
Apr 5
|
We covered Section 4.9, which concerns
Markov chains. After reading this section
you will know the meaning of
"probability vector", "(regular) stochastic matrix",
"Markov chain" and "steady-state vector". You should be able
to find a steady-state vector for a stochastic matrix.
|
Apr 8
|
Read Section 5.1.
We defined eigenvectors, eigenvalues and eigenspaces
for linear transformations and for matrices.
Our intuition is that eigenvectors point in
"preserved directions", and that eigenvalues
measure the amount of "stretching" in a preserved direction.
We explained why λ is an e-value of A iff
A-λI is noninvertible, and why
Vλ =
Nul(A-λI).
We showed that the e-values of a triangular matrix
are exactly the diagonal values.
Quiz 9.
|
Apr 10
|
Read Section 5.2.
We defined the characteristic polynomial
χA(λ) of a matrix A.
We noted that if the entries of A belong to
a number system ℤ, ℚ, ℝ,
ℂ (or any other commutative ring), then the coefficients of
χA(λ) belong to the same number system.
We discussed the structure of the roots of a polynomial
p(λ) of degree n over ℂ versus ℝ.
In particular, a complex polynomial of degree n
always has n complex roots
counting multiplicity, but a real polynomial of degree
n may have AT MOST n real roots (possibly fewer than n).
Consequently, a complex or
real n×n-matrix
has exactly n complex eigenvalues counting multiplicity,
but a real n×n-matrix
has AT MOST n real eigenvalues (possibly fewer).
|
Apr 12
|
Read Section 5.3.
We proved that a transformation T:V→V is diagonalizable iff
V has a basis of e-vectors for T, and that the corresponding
e-values appear on the diagonal, in the proper order,
in the diagonal form for T. We worked through some examples
of diagonalizable and nondiagonalizable transformations.
|
Apr 15
|
We defined A to be similar to B (written A ~ B) if there
is an invertible S such that
A =
S-1BS.
We call S-1BS the conjugate of B by S,
so A ~ B if A is a conjugate of B. We showed that
(i) A ~ A for any square matrix A;
(ii) if A ~ B, then B ~ A;
(iii) if A ~ B ~ C, then A ~ C;
(iv) if A ~ B, then A and B have the same characteristic polynomials
(χA(λ)=χB(λ));
(v) if A ~ B, then A and B have the same e-values;
(vi) if A ~ B (because A =
S-1BS), then
S:
VλA→
VλB
is an isomorphism between the λ-eigenspaces
of A and B for each λ;
(vii) A is diagonalizable iff it is similar to a diagonal matrix.
Quiz 10.
|
Apr 17
|
We defined the sum of subspaces.
We defined when a set of subspaces is independent.
We defined when a sum of subspaces is a direct sum.
We explained why the following are equivalent for a
set {V1,…,Vk} of subspaces:
(i) Each Vi is disjoint from the sum of the others.
(ii) Each vector in the sum is expressible in a unique way
as a sum of vectors from the constituents.
(iii) If you create a set S by
selecting a nonzero vector from each nonzero Vi,
then S is independent.
We ended with the theorem that the e-spaces for a linear
transformation are independent. If the transformation
has a full set of e-vectors, then the whole space
is a direct sum of e-spaces.
|
Apr 19
|
We proved the theorem that the sum of e-spaces of
T is a direct sum, and that T acts diagonally on this subspace.
Then we defined Jordan blocks and Jordan canonical form (JCF).
We stated the theorem that every complex matrix is similar
to a matrix in JCF that is unique up to a permutation
of Jordan blocks. We defined generalized e-spaces and
e-chains, and described through pictures how one
finds the JCF of a matrix.
|
Apr 22
|
We discussed this
handout on diagonalization and JCF. We explained
how to determine the JCF of a matrix using only the nullities
of the matrices (A-λI)k as
λ ranges over the e-values of A
and k ranges over the numbers from 1 to n.
Quiz 11.
|
Apr 24
|
We spent ~20 minutes discussing HW, then turned
to complex numbers. (Read Section 5.5.)
We defined the complex number system. We described the
planar representation of complex numbers and defined
the real and imaginary parts of a complex number, and
the "conjugate", "absolute value", and "argument" of a complex number.
We noted that if A is a real matrix, then its nonreal e-values
come in conjugate pairs. If λ is a nonreal e-value, then
the conjugation map is a real vector space isomorphism
from the e-space Vλ to its conjugate V
λ.
This holds for generalized e-spaces too. Thus we can deal with
complex e-values and e-spaces in conjugate pairs when we eliminate
the use of complex numbers (which shall be described Friday).
|
Apr 26
|
Read Section 6.1.
Do practice problems:
Section 5.5: 24, 25
Chap 5 Supp.: 1, 2, 5, 7, 15, 17
We described how to put real matrices in approximate
diagonal form (or approximate JCF) while avoiding the
use of complex numbers.
We started discussing how to compute length and angle in
ℝn. We defined the dot product and
showed that for vectors in the plane
the length of vector u is ||u||:=
(u•u)1/2, and the angle
θ between vectors u and v satisfies
u•v=||u|| ||v||cos(θ).
We used this formula to show that the angle between two adjacent
diagonals on the face of a cube meet at 60 degrees.
|
Apr 29
|
Read Sections 6.2 and 6.5.
Do practice problems:
Section 6.1: 1, 5, 7, 9, 13, 17, 19, 20
Section 6.2: 1, 3, 17, 23, 24, 27, 29
Section 6.5: 3, 17, 18, 22, 25
We reviewed the relationships between dot product, length
and angle. We mentioned the link between
dot product and matrix multiplication, namely
u•v =
uTv,
which allows us to derive the basic arithmetical laws of dot product
from the arithmetical laws of matrix multiplication.
We discussed how to find a unit vector
in a given direction. Then we turned to a discussion of
orthogonality. We defined
u⊥v ⇔
u•v = 0⇔
uTv = 0.
We defined the orthogonal complement of a set of vectors,
and explained why orthogonal complements are subspaces.
We observed that Nul(A) = Row(A)⊥, and pointed out
that this observation yields an algorithm for finding orthogonal complements.
(Namely, if S is a set of vectors, make them the rows of
a matrix A,
use the nullspace algorithm to find S⊥=Nul(A).)
We stated without proof that in a finite dimensional
space, every subspace equals its double complement.
We defined orthonormal bases, and defined a matrix to be
orthogonal if its columns form an orthonormal basis.
We mentioned that orthogonal matrices are exactly the
matrices of linear transformations that preserve length and angle.
(In fact, orthogonal matrices are exactly those that
preserve dot product. This means that O is orthogonal iff
u•v =
(Ou)•(Ov) for all vectors
u and v.)
A compact definition of orthogonal matrix is:
a matrix O satisfying OTO = I.
Examples of orthogonal matrices include rotation matrices and permutation
matrices.
The second part of the lecture focused on the problem of finding
the best approximate solution to an
inconsistent system. These are systems of the form
Ax=b where b∉Col(A).
A best approximate solution, or least squares solution,
is any solution to the normal equations, which are the
equations derived from Ax=b by left multiplying
by AT: namely,
ATAx=ATb. We explained why
any solution to ATAx=ATb
is a solution to Ax=b̂
where b̂ is the orthogonal projection of
b onto Col(A).
Quiz 12.
|
May 1
|
Read Sections 6.3 and 6.4.
Do practice problems:
Section 6.3: 5, 11, 21, 22
Section 6.4: 9, 17, 18
We worked out a least squares
problem about fitting a curve to a set of data.
Then we developed a formula for
calculating
the orthogonal projection of a vector onto a 1-dimensional subspace.
We used the formula to show how to orthogonalize
a basis for a subspace (Gram-Schmidt).
Finally we extended our earlier formula to one for
calculating
the orthogonal projection of a vector onto an
arbitrary subspace.
I distributed a
review sheet for the final.
|
May 3
|
We reviewed for the final.
(George made another
solutions page for the review sheet. Thanks George!)
|
|
|