The Characteristic Equation#
Today we deepen our study of linear dynamical systems,
systems that evolve according to the equation:
Let’s look at some examples of how such dynamical systems can evolve in
First we’ll look at the system corresponding to:
Next let’s look at the system corresponding to:
Next let’s look at the system corresponding to:
There are very different things happening in these three cases!
Can we find a general method for understanding what is going on in each case?
The study of eigenvalues and eigenvectors is the key to acquiring that understanding.
We will see that to understand each case, we need to learn how to extract the eigenvalues and eigenvectors of
Finding #
In the last lecture we saw that, if we know an eigenvalue
Today we’ll discuss how to determine the eigenvalues of a matrix
The theory will make use of the determinant of a matrix.
Let’s recall that the determinant of a
We also have learned that
(Recall that the inverse of of
Let’s use these facts to help us find the eigenvalues of a
Example.
Let’s find the eigenvalues of
Solution. We must find all scalars
has a nontrivial solution.
By the Invertible Matrix Theorem, this problem is equivalent to finding all
Now,
We know that
So the eigenvalues of
Since
If
The same idea works for
Determinants of Larger Matrices#
Previously, we’ve defined a determinant for a
To find eigenvalues for larger matrices, we need to define the determinant for any sized (ie,
Definition. Let
Then the determinant of
If
If
In other words:
Example. Compute
Solution. The following row reduction uses one row interchange:
So
The remarkable thing is that any other way of computing the echelon form gives the same determinant. For example, this row reduction does not use a row interchange:
Using this echelon form to compute the determinant yields
Invertibility#
The formula for the determinant shows that
We have yet another part to add to the Invertible Matrix Theorem:
Let
The determinant of
is not zero.The number 0 is not an eigenvalue of
.
The second fact is true because if zero is an eigenvalue of
Some facts about determinants:
If
is triangular, then is the product of the entries on the main diagonal of .
The Characteristic Equation#
So,
To return to the question of how to compute eigenvalues of
We capture this fact using the characteristic equation:
We can conclude that
Example. Find the characteristic equation of
Solution. Form
So the characteristic equation is:
Expanding this out we get:
Notice that, once again,
In fact, for any
We say that the eigenvalue 5 in this example has multiplicity 2, because
Example. The characteristic polynomial of a
Solution Factor the polynomial
So the eigenvalues are 0 (with multiplicity 4), 6, and -2.
Since the characteristic polynomial for an
Note that even for a real matrix, eigenvalues may sometimes be complex.
Practical Issues.
These facts show that there is, in principle, a way to find eigenvalues of any matrix. However, you need not compute eigenvalues for matrices larger than
Similarity#
An important concept for things that come later is the notion of similar matrices.
Definition. If
Similarity is symmetric, so if
Changing
An important way to think of similarity between
Theorem. If
Proof. If
Now let’s construct the characteristic polynomial by taking the determinant:
Using the properties of determinants we discussed earlier, we compute:
Since
Markov Chains#
Let’s return to the problem of solving a Markov Chain.
At this point, we can place the theory of Markov Chains into the broader context of eigenvalues and eigenvectors.
Theorem. The largest eigenvalue of a Markov Chain is 1.
Proof. First of all, it is obvious that 1 is an eigenvalue of a Markov chain since we know that every Markov Chain
To prove that 1 is the largest eigenvalue, recall that each column of a Markov Chain sums to 1.
So, every row of
Hence, for any vector
So
Now, any matrix
To see this, suppose
But det(
So then
So there can be no
The Complete Solution for a Markov Chain#
Previously, we were only able to ask about the “eventual” steady state of a Markov Chain.
But another crucial question is: how quickly does a particular Markov Chain reach steady state from some initial starting condition?
By completely solving a Markov Chain, we will determine both its steady state and its rate of convergence.
Let’s use an example: we previously studied the Markov Chain defined by
Let’s ask how quickly it reaches steady state, from the starting point defined as
Remember that

Using the methods we studied today, we can find the characteristic equation:
Using the quadratic formula, we find the roots of this equation to be 1 and 0.92.
(Note that, as expected, the largest eigenvalue of the chain is 1.)
Next, using the methods in the previous lecture, we find a basis for each eigenspace of
For
For
Next, we write
To write
In other words:
The matrix
So, now we can put it all together.
We know that
and we know the values of
So let’s compute each
Now note the power of the eigenvalue approach:
And so in general:
And using the
This explicit formula for
In other words:
The equation tells us how quickly the chain converges: like
Thus
