Suggestions: After you read 8.5, try the first set of problems
from 8.5. Then read 7.1 (and the tail of 8.5) and do the problems
from 7.1. Lastly, read the rest of 8.5, read 7.2 and do the problems
for 7.2 and the last couple problems from 8.5.
Print this page and keep it handy, for
the notes given below, as you do your reading and work on the
problems. (Scroll down to see.)
Reading
- Read 8.5 (Similarity) up to the discussion of eigenvalues that
starts at the bottom of p408.
- Then read 7.1. (Eigenvalues, Eigenvectors - for an n x n
matrix A). Look back to Exple 6 p100 at the point where it's
suggested to do so.
- Then, at p408 in 8.5, read Exple 4 and the paragraph or so
that precede(s) it. This example has to do with eigenvalues and
eigenvectors for a linear transformation T (which match up with
eigenvalues and eigenvectors for a representing matrix for
T).
- Read 7.2 (Diagonalization).
Problems
- 8.5 p411: 2, 5, 6, 7, 9c
Click to view
solutions
- 7.1 p344: 1a-2a-3a, 1e-2e-3e, 4a-5a-6a, 10ac
Click to view
solutions
- 7.2 p354: 3, 5, 13, 15
Click to view
solutions
- 8.5 p411: 13
Click to view
solutions
Go back to the top
Notes to summarize the reading/classnotes for
8.5.
- The transition matrix for changing from one basis to another
is the same as the matrix that would be used to carry out the
identity transformation using the old basis for inputs and the new
basis for outputs. In other words, if B = {u1,
u2, ..., un} is one basis for
some n-dimensional vector space V, and B' is another basis for V,
then the transition matrix used to change from B-coordinates to B'
coordinates is the n x n matrix whose jth column contains the B'
coordinates for the original jth basis vector
uj. In other words, this matrix, denoted
[I]B,B' in the text, is
[I]B,B' = [
[u1]B' |
[u2]B' | ...|
[un]B' ].
- Question: Given two different bases for the same n-dimensional
vector space V, and given some linear operator T:V-->V, how are
the matrix that represents T relative to one basis and the matrix
that represents T relative to the other basis related?
- Why do we care? If we can understand the answer, then we
may be able to choose a good basis. Good basis in this case
would mean one that makes the representing matrix as simple as
possible, for instance a diagonal matrix.
- Answer: Say that the two bases are B and B'. Denote the two
representing matrices by [T]B and
[T]B', respectively. Let P be the transition
matrix from B' to B. Let Q = P-1 = the transition
matrix from B to B'.Then
[T]B' = P-1
[T]B P which says the same as
[T]B' = Q [T]B P
[T]B = Q-1
[T]B' Q which says the same as
[T]B = P [T]B' Q
- Why this answer? Think of B as a language, B' as another
language. Given the B'-language version of a vector v,
one way to find the B'-language version of T(v) is
simply to multiply by the matrix [T]B'.
Another way to do the same thing would be to (1) change
languages (2) carry out the transformation T in the new
language (3) change languages back. In matrix form this would
mean: given the B'-language version of v, we would (1)
first multiply by P; (2) then multiply by
[T]B; (3) then multiply by Q. Thus
multiplication by [T]B' and multiplication
by Q [T]B P both accomplish exactly the same
thing. In other words both have exactly the same effect on all
the x's in Rn. This gives us that they
must actually be the same matrix.
- See Exple 1 p405
- Definition. If A and B are square matrices (of the same size),
then we say that B is similar to A iff there is an
invertible matrix P such that B = P-1 A P.
- It is pretty easy to show that B similar to A ==> A
similar to B.
- So usually we just say A and B are similar.
- The question and answer just above could be restated as
follows: all of the matrices that could be used to represent
some linear operator T:V-->V (where V is some n-dimensional
vector space) are similar to each other.
- When two matrices are similar, they have a lot in common!
See p 406.
Notes to guide the reading for
7.1.
Go back to the top
Notes to guide the reading for 7.2
- This section begins with two related problems concerning a
square (n x n, say) matrix A.
- Eigenvector Problem for A: Does Rn have a
basis consisting entirely of eigenvectors for A?
- Diagonalization Problem for A is Does there exist a
diagonal matrix D that is similar to A? In other words, do
there exist a diagonal matrix D and an invertible matrix P such
that D = P-1AP?
- Why do these problems matter? See "Why do eigenvectors
matter?" above. They are driven by the problem of trying to find a
basis that will let us represent a linear transformation by a
diagonal matrix (a nice matrix!).
- Theorem connecting the two problems.
- Theorem. For a given A, either the "Eigenvector Problem"
and the "Diagonalization Problem" both have a solution or
neither has a solution. (i.e. both answers are yes or both
answers are no).
- More precisely, for a given A, whether the Diagonalization
Problem can be solved depends on whether one can find enough
independent eigenvectors for A. How many is enough? Enough to
make a basis for Rn.
- So how do we find eigenvectors? That's what's covered in
7.1.
- If we do find enough independent eigenvectors how do we use
them to diagonalize A? See the procedure outlined on p349. See
Exple 1 p349.
- Shortcuts for knowing we have (enough) independent
eigenvectors.
- Thm 7.2.3 If A has n distinct real eigenvalues, then it
will be possible to diagonalize A.
- Why? See Thm 7.2.2 below.
- What if not? When an an x n matrix A fails to have n
distinct real eigenvalues then it might or might not be
possible to diagonalize, with the deciding factor still
being whether or not there are n independent eigenvectors.
See Exple 1 p349 and Exple 2 p340.
- Thm 7.2.2 If A has n distinct real eigenvalues and we
choose one eigenvector for each eigenvalue, then the resulting
set of eigenvectors will be independent (so there will be
enough). That's why it will be possible to diagonalize A in
this case. See Exple 3 p352, Exple 4 p352.
- 7.2.2 is a little more general.
- Whenever we have distinct eigenvalues, however many, if
we take one eigenvector for each eigenvalue, then the
corresponding set of eigenvectors will be independent.
- What 7.2.2. does not say: if chosen eigenvectors do not
correspond to distinct eigenvalues then the eigenvectors
won't form an independent set. Sometimes they do. (Sometimes
they don't.) See Exple 1 p349 and Exple 2 p350.
- A more general theorem. Given several distinct eigenvalues,
if we find a basis for each of the corresponding eigenspaces
and then "merge" these bases into a single set of eigenvectors,
then the resulting set of eigenvectors will always be
independent. See the Remark on p352.
- For instance, say that the (only) eigenvalues are 1 and
2, and that the eigenspace corresponding to the eigenvalue 1
has dimension 2 and the eigenspace corresponding to the
eigenvalue 2 has dimension 3. Then if we find a basis for
the first eigenspace (so 2 vectors) and find a basis for the
second eigenspace (so 3 vectors) and merge these two basis
(so 5 vectors all together) then the 5 chosen eigenvectors
form an independent set. If the matrix A happened to be 5 x
5 then we can diagonalize. But if A is bigger than that we
can't.
- Thm 7.2.4. Yet another criterion for diagonalizability:
- A is diagonalizable iff, for each eigenvalue
l0, the dimension of
the corresponding eigenspace is the same as the number of
times that (l-l0)
appears as a factor of the characteristic polynomial.
- The former is always less than or equal the latter. The
criterion requires it be equal.
- For instance, if the characteristic polynomial is
(l-1)2(l-2)3
then the eigenspace corresponding
to the eigenvalue 1 should have dimension 2 (it might only
have dimension 1) and the eigenspace corresponding to the
eigenvalue 2 should have dimension 3 (it might only have
dimension 1 or 2).
- See the alternative solution for
Exple 2 p350.
- See p353 "Geometric and Algebraic
Multiplicity", if you like, for some alternative language
that is used in this connection.
- Iterative processes often involve using
higher and higher powers of the same matrix. Computing these
powers can be a pain. If the matrix can be diagonalized, computing
these powers becomes much, much easier. See pp353-354, especially
Exple 5 and the remark that follows it.
Go back to the top
- Alexia Sontag,
Mathematics
- Wellesley College
- Date Created: January 4, 2001
- Last Modified: May 3, 2002
- Expires: June 30, 2002