Home

Syllabus

Lecture Topics

Homework

Test Solutions

Policies


Math 3130: Introduction to Linear Algebra, Spring 2013


Lecture Topics


Date
What we discussed/How we spent our time
Jan 14
Syllabus. Text.

Section 1.1: Systems of linear equations. Augmented matrix of a linear system. Elementary row operations.

Jan 16
Section 1.2: Row reduction. (Reduced) row echelon form. Pivots, pivot positions and pivot columns. Free and pivot variables. Solution sets.
Jan 18
We reviewed terminology from Sections 1.1 and 1.2. (row reduction/Gaussian elimination, row echelon form, reduced row echelon form, augmented matrix, coefficient matrix, consistent vs. inconsistent system, existence and uniqueness of solutions, overdetermined and underdetermined systems). Then we began discussing matrix arithmetic (addition, zero matrix additive inverse, multiplication, identity matrix, multiplicative inverses) to prepare to discuss the matrix point of view for solving linear systems, described in Sections 1.3 and 1.4.
Jan 23
We discussed matrix arithmetic, including the fact that the collection of nxn real matrices forms a ring (which is noncommutative when n is greater than 1). We explained how linear systems are expressible as matrix equations of the form Ax = b. Quiz 1.

Read Sections 1.3 and 1.4, and skim Sections 2.1 and 2.2.

Jan 25
We explained that row reduction corresponds to left multiplication by certain invertible matrices. We defined elementary matrices Eii(r), Eji(r), Pij, and general permutation matrices. We explained how to find the inverse of each elementary matrix, and how to find the inverse of a product of invertible matrices.
Jan 28
We defined real vector spaces, and gave geometric representations for vectors and for vector arithmetic (adding and scaling) in ℝn. We defined "linear combination" and "span". We recast the problem of solving a linear system as solving a vector equation and also as the problem of finding the weights needed to express a given vector as linear combination of others. Quiz 2.

Jan 30
We discussed the connection between linear systems, matrix equations and vector equations. We gave informal definitions of "vector space" and "subspace".
Feb 1
Read Section 1.5.

We showed that the solution set of a homogeneous system with n unknowns is a subspace of ℝn, called the nullspace of A, and that the set of all vectors b such that Ax=b is consistent is a subspace of ℝm, called the column space of A. We showed that the general solution to Ax=b may be expressed in terms of the general solution to the associated homogeneous system, Ax=0, along with a particular solution p to Ax=b:

x = p + (t1*s1+ …+tk*sk).

The purpose of this representation is to show that the general solution of a linear system may be viewed as a translate of a subspace.

Feb 4
Read Section 1.7.

We defined linear (in)dependence. The set of columns of a matrix A are linearly independent iff Ax=0 has only the trivial solution. A subset of a linearly independent set is linearly independent. The empty set of vectors is linearly independent. A 1-element set of vectors, {v}, is linearly independent iff v is nonzero. A 2-element set of vectors is linearly independent iff neither vector is a multiple of the other. An n-element set of vectors is linearly independent iff no one is a linear combination of the others. ℝn does not contain a linearly independent set of size greater than n, but does contain a linearly independent set of size m for any m satisfying 0≤m≤n. Quiz 3.

Feb 6
Read Sections 1.8-1.9.

I defined "linear transformation". (Alternative terminology: "vector space homomorphism".) I gave examples and nonexamples. I showed that every linear transformation T:ℝn→ ℝm has the form T(x) = Ax for some uniquely determined (m×n) matrix A, which we showed how to calculate. We examined some examples. One nice example (which I think is worth memorizing) is that the counterclockwise rotation of the plane through an angle θ is a linear transformation from ℝ2 to ℝ2 whose matrix is: rotation matrix.

I explained why a linear transformation T(x) = Ax is 1-1 iff Ax = 0 has only the trivial solution. As class ended, I posed the problem of identifying what property of the matrix A corresponds to the property that the transformation T is onto.

Feb 8
We worked on practice problems. (We discussed the answers to the first four problems in class.)
Feb 11
Read Sections 1.6, 1.10

Introduction to Mathematica: I discussed how to get Mathematica and how to start creating a notebook. Here is a good video tutorial series.

I worked through a simple example using Mathematica to balance a chemical equation. I discussed using linear algebra to solve network flow problems, and to study predator-prey dynamics. Quiz 4.

Feb 13
Reread Sections 2.1-2.2

We discussed the solutions to Quiz 4, especially what it means to "define" a term or phrase. Then we discussed the transpose of a matrix, and the laws of arithmetic related to transpose. The most important (or unexpected) law is that (AB)t=BtAt. (Note: (AB)t≠AtBt, usually. In fact, the dimensions of the matrices usually don't match up properly for this to even make sense.) Finally we discussed left, right and 2-sided inverses of a matrix. We showed that if A has a left inverse, then Ax=0 has only the trivial solution. More about this on Friday …

Feb 15
We practiced writing proofs using this handout. Our practice was prefaced by a brief discussion of (i) how to prove the equivalence of a sequence of statements, and (ii) proof by contradiction.
Feb 18
Read Section 2.3

We continued our discussion of inverses, focusing on the geometric interpretation of "A is left invertible" and "A is right invertible". We added to earlier characterizations of left/right invertibility by proving that A is left invertible iff T(x)=Ax preserves independent sets, and A is right invertible iff T(x)=Ax preserves spanning sets. We observed that a left invertible matrix is right invertible iff it is square. We showed that if A is square, then Gaussian elimination applied to [A|I] yields [I|A-1]. Quiz 5.

Feb 20
Read Sections 2.4-2.5

We first discussed the arithmetic of partitioned matrices. (Everything works as expected, with the obvious restrictions.) We focused briefly on block diagonal matrices and block upper triangular matrices, noting that a matrix of either type is invertible iff its diagonal blocks are invertible.

Second, we discussed the fact that Gaussian elimination applied to a matrix A produces a factorization A=LR where L is a square invertible matrix and R is in echelon form. If Gaussian elimination can be performed without row switches or row scaling, then L is lower triangular with 1's on the diagonal. If A is square, then R is upper triangular (in which case we call it U instead of R). Thus, if Gaussian elimination applied to a square matrix A without row switches or row scalings succeeds, then it produces a factorization A=LU with
(i) L lower triangular with 1's on the diagonal, and
(ii) U upper triangular.

We observed that if A=LU, then the single system Ax=b is equivalent to a pair of simpler systems Ly=b and Ux=y. We mentioned that if A is n×n, then it takes O(n3) arithmetic operations to solve Ax=b for a given b via Gaussian elimination. It takes O(n3) arithmetic operations to factor A as LU. But it takes only O(n2) arithmetic operations to solve each of Ly=b and Ux=y. Thus, if you want to solve many systems of the form Ax=b for the same A but different b's, then it can be more efficient to factor A first.

Feb 22
We continued our discussion of LU factorizations, and saw by example how to find one. We observed that the inverse of a (unit) lower triangular matrix is a (unit) lower triangular matrix, and the product of two (unit) lower triangular matrices is a (unit) lower triangular matrix.

We introduced the Leontief economic model, x=Cx+d, which depicts relationships between different sectors of an economy.

Feb 25
We discussed the relationship between fixed point problems (x=F(x) for some function F) and problems about finding roots (0=G(x) for some function G). In particular, the equation x=Cx+d from the Leontief model can be treated as an ordinary linear system (I-C)x=d , in which case the solution is x=(I-C)-1d, or may be treated as a fixed point problem x=F(x) for F(x)= Cx+d, in which case the solution obtained by iteration is (I+C+C2 +…)d. These solutions are equal, since (I-C)-1=(I+C+C2 +…) when C is "sufficiently small". Here the fact that the entries of C are nonnegative and the column sums of C are less than 1 make C "sufficiently small". Quiz 6.
Feb 27
Midterm review. (George made a solutions page.)
Mar 1
Midterm.
Mar 4
Read Section 2.7.

We discussed affine transformations, and how to represent them in homogeneous coordinates.

Mar 6
Read Sections 2.7-2.9.

We discussed subspaces of ℝn, Col(A), Row(A), Nul(A), ordered bases for a space, standard basis for ℝn, dimension, rank and nullity of a matrix. We gave algorithms for finding bases for Col(A), Row(A), Nul(A).

Mar 8
Read Sections 3.1-3.2.

We proved the rank + nullity theorem and that dimension is well defined. We worked on some practice problems.

Mar 11
We discussed the definition of the determinant, its use in caculating signed volume, and the Laplace expansion. Quiz 7.
Mar 13
We discussed properties of the determinant including: the determinant is defined only for square matrices, the determinant of a matrix is nonzero iff its columns are independent, det(A)=det(AT), and A*adj(A) = det(A)*I (hence A-1=(1/det(A))*adj(A) if A is invertible).
Mar 15
Read Section 3.3.

We discussed properties of the determinant including: the determinant of (block) triangular matrices, determinant of a product is the product of determinants, the determinant can be computed via Gaussian elimination, if T(x)=Ax, then the determinant of A measures the "volume expansion" associated with T, and that the determinant is the unique alternating multilinear function d of n variables defined on ℝn for which d(e1,…,en)=1. We briefly discussed Cramer's Rule.

Mar 18
We worked out a 3×3 example illustrating Cramer's Rule, then discussed the permutation expension of the determinant. Quiz 8.
Mar 20
Read Sections 4.1-4.4.

We discussed the word "abstract", then defined (real) vector spaces. We discussed the new examples Mm×n(ℝ), Pn(t), and Ck([0,1]). We computed that Mm×n(ℝ) has dimension mn, Pn(t) has dimension n+1, and C0([0,1]) is infinite dimensional.

Mar 22
Read Sections 4.5-4.6.

We discussed coordinates relative to a basis, and found coordinates for certain vectors in M2×2(ℝ) and P3(t). We defined the (very important!) terms isomorphism and isomorphic for real vector spaces. We proved the (very important!) theorem that any finitely generated real vector space is isomorphic to ℝn for some n.

Apr 1
Read Section 4.7.

We reviewed the proof that any finitely generated real vector space is isomorphic to ℝn for some n. We then explained why, if V and W are finitely generated vector spaces with ordered bases B and C respectively and T:V→W is a linear tranformation, then there is a unique matrix C[T]B such that

C[T]B [v]B=[T(v)]C.

We calculated this matrix for some examples, namely the matrix for the transpose operation from M2×2(ℝ) to itself and the matrix for the derivative operation from Pn(t) to Pn-1(t). We mentioned that if I:V→V is the identity transformation, then C[I]B is the change of basis matrix from basis B to basis C.

Apr 3
We discussed change of basis a bit more, then worked on these practice problems.
Apr 5
We covered Section 4.9, which concerns Markov chains. After reading this section you will know the meaning of "probability vector", "(regular) stochastic matrix", "Markov chain" and "steady-state vector". You should be able to find a steady-state vector for a stochastic matrix.
Apr 8
Read Section 5.1.

We defined eigenvectors, eigenvalues and eigenspaces for linear transformations and for matrices. Our intuition is that eigenvectors point in "preserved directions", and that eigenvalues measure the amount of "stretching" in a preserved direction. We explained why λ is an e-value of A iff A-λI is noninvertible, and why Vλ = Nul(A-λI). We showed that the e-values of a triangular matrix are exactly the diagonal values. Quiz 9.

Apr 10
Read Section 5.2.

We defined the characteristic polynomial χA(λ) of a matrix A. We noted that if the entries of A belong to a number system ℤ, ℚ, ℝ, ℂ (or any other commutative ring), then the coefficients of χA(λ) belong to the same number system.

We discussed the structure of the roots of a polynomial p(λ) of degree n over ℂ versus ℝ. In particular, a complex polynomial of degree n always has n complex roots counting multiplicity, but a real polynomial of degree n may have AT MOST n real roots (possibly fewer than n). Consequently, a complex or real n×n-matrix has exactly n complex eigenvalues counting multiplicity, but a real n×n-matrix has AT MOST n real eigenvalues (possibly fewer).

Apr 12
Read Section 5.3.

We proved that a transformation T:V→V is diagonalizable iff V has a basis of e-vectors for T, and that the corresponding e-values appear on the diagonal, in the proper order, in the diagonal form for T. We worked through some examples of diagonalizable and nondiagonalizable transformations.

Apr 15
We defined A to be similar to B (written A ~ B) if there is an invertible S such that A = S-1BS. We call S-1BS the conjugate of B by S, so A ~ B if A is a conjugate of B. We showed that
(i) A ~ A for any square matrix A;
(ii) if A ~ B, then B ~ A;
(iii) if A ~ B ~ C, then A ~ C;
(iv) if A ~ B, then A and B have the same characteristic polynomials (χA(λ)=χB(λ));
(v) if A ~ B, then A and B have the same e-values;
(vi) if A ~ B (because A = S-1BS), then S: VλA→ VλB is an isomorphism between the λ-eigenspaces of A and B for each λ;
(vii) A is diagonalizable iff it is similar to a diagonal matrix.
Quiz 10.
Apr 17
We defined the sum of subspaces. We defined when a set of subspaces is independent. We defined when a sum of subspaces is a direct sum. We explained why the following are equivalent for a set {V1,…,Vk} of subspaces:
(i) Each Vi is disjoint from the sum of the others.
(ii) Each vector in the sum is expressible in a unique way as a sum of vectors from the constituents.
(iii) If you create a set S by selecting a nonzero vector from each nonzero Vi, then S is independent.
We ended with the theorem that the e-spaces for a linear transformation are independent. If the transformation has a full set of e-vectors, then the whole space is a direct sum of e-spaces.
Apr 19
We proved the theorem that the sum of e-spaces of T is a direct sum, and that T acts diagonally on this subspace. Then we defined Jordan blocks and Jordan canonical form (JCF). We stated the theorem that every complex matrix is similar to a matrix in JCF that is unique up to a permutation of Jordan blocks. We defined generalized e-spaces and e-chains, and described through pictures how one finds the JCF of a matrix.
Apr 22
We discussed this handout on diagonalization and JCF. We explained how to determine the JCF of a matrix using only the nullities of the matrices (A-λI)k as λ ranges over the e-values of A and k ranges over the numbers from 1 to n. Quiz 11.
Apr 24
We spent ~20 minutes discussing HW, then turned to complex numbers. (Read Section 5.5.)

We defined the complex number system. We described the planar representation of complex numbers and defined the real and imaginary parts of a complex number, and the "conjugate", "absolute value", and "argument" of a complex number. We noted that if A is a real matrix, then its nonreal e-values come in conjugate pairs. If λ is a nonreal e-value, then the conjugation map is a real vector space isomorphism from the e-space Vλ to its conjugate V λ. This holds for generalized e-spaces too. Thus we can deal with complex e-values and e-spaces in conjugate pairs when we eliminate the use of complex numbers (which shall be described Friday).

Apr 26
Read Section 6.1.

Do practice problems:
Section 5.5: 24, 25
Chap 5 Supp.: 1, 2, 5, 7, 15, 17

We described how to put real matrices in approximate diagonal form (or approximate JCF) while avoiding the use of complex numbers.

We started discussing how to compute length and angle in ℝn. We defined the dot product and showed that for vectors in the plane the length of vector u is ||u||:= (u•u)1/2, and the angle θ between vectors u and v satisfies u•v=||u|| ||v||cos(θ). We used this formula to show that the angle between two adjacent diagonals on the face of a cube meet at 60 degrees.

Apr 29
Read Sections 6.2 and 6.5.

Do practice problems:
Section 6.1: 1, 5, 7, 9, 13, 17, 19, 20
Section 6.2: 1, 3, 17, 23, 24, 27, 29
Section 6.5: 3, 17, 18, 22, 25

We reviewed the relationships between dot product, length and angle. We mentioned the link between dot product and matrix multiplication, namely uv = uTv, which allows us to derive the basic arithmetical laws of dot product from the arithmetical laws of matrix multiplication. We discussed how to find a unit vector in a given direction. Then we turned to a discussion of orthogonality. We defined uvuv = 0uTv = 0. We defined the orthogonal complement of a set of vectors, and explained why orthogonal complements are subspaces. We observed that Nul(A) = Row(A), and pointed out that this observation yields an algorithm for finding orthogonal complements. (Namely, if S is a set of vectors, make them the rows of a matrix A, use the nullspace algorithm to find S=Nul(A).) We stated without proof that in a finite dimensional space, every subspace equals its double complement. We defined orthonormal bases, and defined a matrix to be orthogonal if its columns form an orthonormal basis. We mentioned that orthogonal matrices are exactly the matrices of linear transformations that preserve length and angle. (In fact, orthogonal matrices are exactly those that preserve dot product. This means that O is orthogonal iff uv = (Ou)•(Ov) for all vectors u and v.) A compact definition of orthogonal matrix is: a matrix O satisfying OTO = I. Examples of orthogonal matrices include rotation matrices and permutation matrices.

The second part of the lecture focused on the problem of finding the best approximate solution to an inconsistent system. These are systems of the form Ax=b where b∉Col(A). A best approximate solution, or least squares solution, is any solution to the normal equations, which are the equations derived from Ax=b by left multiplying by AT: namely, ATAx=ATb. We explained why any solution to ATAx=ATb is a solution to Ax=b̂ where b̂ is the orthogonal projection of b onto Col(A). Quiz 12.

May 1
Read Sections 6.3 and 6.4.

Do practice problems:
Section 6.3: 5, 11, 21, 22
Section 6.4: 9, 17, 18

We worked out a least squares problem about fitting a curve to a set of data. Then we developed a formula for calculating the orthogonal projection of a vector onto a 1-dimensional subspace. We used the formula to show how to orthogonalize a basis for a subspace (Gram-Schmidt). Finally we extended our earlier formula to one for calculating the orthogonal projection of a vector onto an arbitrary subspace. I distributed a review sheet for the final.

May 3
We reviewed for the final. (George made another solutions page for the review sheet. Thanks George!)