By the end of this section, you should be able to give precise and thorough answers to the questions listed below. You may want to keep these questions in mind to focus your thoughts as you complete the section.
What is a diagonal matrix?
What does it mean to diagonalize a matrix?
What does it mean for two matrices to be similar?
What important properties do similar matrices share?
Under what conditions is a matrix diagonalizable?
When a matrix is diagonalizable, what is the structure of a matrix that diagonalizes ?
In 1202 Leonardo of Pisa (better known as Fibonacci) published Liber Abaci (roughly translated as The Book of Calculation), in which he constructed a mathematical model of the growth of a rabbit population. The problem Fibonacci considered is that of determining the number of pairs of rabbits produced in a given time period beginning with an initial pair of rabbits. Fibonacci made the assumptions that each pair of rabbits more than one month old produces a new pair of rabbits each month, and that no rabbits die. (We ignore any issues about that might arise concerning the gender of the offspring.) If we let represent the number of rabbits in month , Fibonacci produced the model
is a very famous sequence in mathematics and is called the Fibonacci sequence. This sequence is thought to model many natural phenomena such as number of seeds in a sunflower and anything which grows in a spiral form. It is so famous in fact that it has a journal devoted entirely to it. As a note, while Fibonacci's work Liber Abaci introduced this sequence to the western world, it had been described earlier Sanskrit texts going back as early as the sixth century.
By definition, the Fibonacci numbers are calculated by recursion. This is a very ineffective way to determine entries for large . Later in this section we will derive a fascinating and unexpected formula for the Fibonacci numbers using the idea of diagonalization.
As we have seen when studying Markov processes, each state is dependent on the previous state. If is the initial state and is the transition matrix, then the th state is found by . In these situations, and others, it is valuable to be able to quickly and easily calculate powers of a matrix. We explore a way to do that in this section.
Consider a very simplified weather forecast. Let us assume there are two possible states for the weather: rainy () or sunny(). Let us also assume that the weather patterns are stable enough that we can reasonably predict the weather tomorrow based on the weather today. If is is sunny today, then there is a 70% chance that it will be sunny tomorrow, and if it is rainy today then there is a 40% chance that it will be rainy tomorrow. If is a state vector that indicates a probability that it is sunny and probability that it is rainy on day , then
tells us the likelihood of it being sunny or rainy on day 1. Let .
The previous result demonstrates that to determine the long-term probability of a sunny or rainy day, we want to be able to easily calculate powers of the matrix . Use a computer algebra system (e.g., Maple, Mathematica, WolframAlpha) to calculate the entries of ,, and . Based on this data, what do you expect the long term probability of any day being a sunny one?
In Preview Activity 19.1 we saw how if we can powers of a matrix we can make predictions about the long-term behavior of some systems. In general, calculating powers of a matrix can be a very difficult thing, but there are times when the process is straightforward.
Activity 19.2 illustrates that calculating powers of square matrices whose only nonzero entries are along the diagonal is rather simple. In general, if
for any positive integer . Recall that a diagonal matrix is a matrix whose only nonzero elements are along the diagonal (see Definition 8.6). In this section we will see that matrices that are similar to diagonal matrices have some very nice properties, and that diagonal matrices are useful in calculations of powers of matrices.
As Activity 19.3 illustrates, to calculate the powers of a matrix of the form we only need determine the powers of the matrix . If is a diagonal matrix, this is especially straightforward.
Similar matrices play an important role in certain calculations. For example, Activity 19.3 showed that if we can write a square matrix in the form for some invertible matrix and diagonal matrix , then finding the powers of is straightforward. As we will see, the relation will imply that the matrices and share many properties.
Activity 19.4 suggests that similar matrices share some, but not all, properties. Note that if , then with . So if is similar to , then is similar to . Similarly (no pun intended), since (where is the identity matrix), then any square matrix is similar to itself. Also, if and , then . So if is similar to and is similar to , then is similar to . If you have studied relations, these three properties show that similarity is an equivalence relation on the set of all matrices. This is one reason why similar matrices share many important traits, as the next activity highlights.
When a matrix is similar to a diagonal matrix, we can gain insight into the action of the corresponding matrix transformation. As an example, consider the matrix transformation from to defined by , where
We are interested in understanding what this matrix transformation does to vectors in . First we note that has eigenvalues and with corresponding eigenvectors and . If we let , then you can check that
So transforms the standard coordinate system into a coordinate system in which columns of determine the axes, as illustrated in the second picture in Figure 19.3. Applying to the output scales by 2 in the first component and by in the second component as depicted in the third picture in Figure 19.3. Finally, we apply to translate back into the standard coordinate system as shown in the last picture in Figure 19.3. In this case, we can visualize that when we apply the transformation to a vector in this system it is just scaled in the system by the matrix . Then the matrix translates everything back to the standard coordinate system.
This geometric perspective provides another example of how having a matrix similar to a diagonal matrix informs us about the situation. In what follows we determine the conditions that determine when a matrix is similar to a diagonal matrix.
In Preview Activity 19.1 and in the matrix transformation example we found that a matrix was similar to a diagonal matrix whose columns were eigenvectors of . This will work for a general matrix as long as we can find an invertible matrix whose columns are eigenvectors of . More specifically, suppose is an matrix with linearly independent eigenvectors ,,, with corresponding eigenvalues ,,, (not necessarily distinct). Let
The key notion to the process described above is that in order to diagonalize an matrix , we have to find linearly independent eigenvectors for . When is diagonalizable, a matrix so that is diagonal is said to diagonalize .
It should be noted that there are square matrices that are not diagonalizable. For example, the matrix has 1 as its only eigenvalue and the dimension of the eigenspace of corresponding to the eigenvalue is one. Therefore, it will be impossible to find two linearly independent eigenvectors for .
We showed previously that eigenvectors corresponding to distinct eigenvalue are always linearly independent, so if an matrix has distinct eigenvalues then is diagonalizable. Activity 19.6 (b) shows that it is possible to diagonalize an matrix even if the matrix does not have distinct eigenvalues. In general, we can diagonalize a matrix as long as the dimension of each eigenspace is equal to the multiplicity of the corresponding eigenvalue. In other words, a matrix is diagonalizable if the geometric multiplicity is the same is the algebraic multiplicity for each eigenvalue.
At this point we might ask one final question. We argued that if an matrix has linearly independent eigenvectors, then is diagonalizable. It is reasonable to wonder if the converse is true โ that is, if is diagonalizable, must have linearly independent eigenvectors? The answer is yes, and you are asked to show this in Exercise 6. We summarize the result in the following theorem.
An matrix is diagonalizable if and only if has linearly independent eigenvectors. If is diagonalizable and has linearly independent eigenvectors ,,, with for each , then matrix whose columns are linearly independent eigenvectors of satisfies , where is the diagonal matrix with diagonal entries for each .
Determine if is diagonalizable. If diagonalizable, find a matrix that diagonalizes .
Solution.
Technology shows that the characteristic polynomial of is
.
The eigenvalues of are the solutions to the characteristic equation . Thus, the eigenvalues of are and . To find a basis for the eigenspace of corresponding to the eigenvalue , we find the general solution to the homogeneous system . Using technology we see that the reduced row echelon form of is . So if , then the general solution to is
.
So a basis for the eigenspace of corresponding to the eigenvalue is
.
To find a basis for the eigenspace of corresponding to the eigenvalue , we find the general solution to the homogeneous system . Using technology we see that the reduced row echelon form of is . So if , then the general solution to is
.
So a basis for the eigenspace of corresponding to the eigenvalue is
.
Eigenvectors corresponding to different eigenvalues are linearly independent, so the set
is a basis for . Since we can find a basis for consisting of eigenvectors of , we conclude that is diagonalizable. Letting
Determine if is diagonalizable. If diagonalizable, find a matrix that diagonalizes .
Solution.
Technology shows that the characteristic polynomial of is
.
The eigenvalues of are the solutions to the characteristic equation . Thus, the eigenvalues of are and . To find a basis for the eigenspace of corresponding to the eigenvalue , we find the general solution to the homogeneous system . Using technology we see that the reduced row echelon form of is . So if , then the general solution to is
.
So a basis for the eigenspace of corresponding to the eigenvalue is
.
To find a basis for the eigenspace of corresponding to the eigenvalue , we find the general solution to the homogeneous system . Using technology we see that the reduced row echelon form of is . So if , then the general solution to is
.
So a basis for the eigenspace of corresponding to the eigenvalue is
.
Since each eigenspace is one-dimensional, we cannot find a basis for consisting of eigenvectors of . We conclude that is not diagonalizable.
Is it possible for two matrices and to have the same eigenvalues with the same algebraic multiplicities, but one matrix is diagonalizable and the other is not? Explain.
Solution.
Yes it is possible for two matrices and to have the same eigenvalues with the same multiplicities, but one matrix is diagonalizable and the other is not. An example is given by the matrices and in this problem.
Is it possible to find diagonalizable matrices and such that is not diagonalizable? If yes, provide an example. If no, explain why.
Solution.
Let and . Since and are both diagonal matrices, their eigenvalues are their diagonal entries. With distinct eigenvalues, both and are diagonalizable. In this case we have , whose only eigenvector is . The reduced row echelon form of is . So a basis for the eigenspace of is . Since there is no basis for consisting of eigenvectors of , we conclude that is not diagonalizable.
Is it possible to find diagonalizable matrices and such that is not diagonalizable? If yes, provide an example. If no, explain why.
Solution.
Let and . Since and are both diagonal matrices, their eigenvalues are their diagonal entries. With distinct eigenvalues, both and are diagonalizable. In this case we have , whose only eigenvector is . The reduced row echelon form of is . So a basis for the eigenspace of is . Since there is no basis for consisting of eigenvectors of , we conclude that is not diagonalizable.
Is it possible to find a diagonalizable matrix such that is not diagonalizable? If yes, provide an example. If no, explain why.
Solution.
It is not possible to find a diagonalizable matrix such that is not diagonalizable. To see why, suppose that matrix is diagonalizable. That is, there exists a matrix such that , where is a diagonal matrix. Recall that . So
Is it possible to find an invertible diagonalizable matrix such that is not diagonalizable? If yes, provide an example. If no, explain why.
Solution.
It is not possible to find an invertible diagonalizable matrix such that is not diagonalizable. To see why, suppose that matrix is diagonalizable. That is, there exists a matrix such that , where is a diagonal matrix. Thus, . Since is invertible, . It follows that . So none of the diagonal entries of can be . Thus, is invertible and is a diagonal matrix. Then
A matrix is diagonalizable if there is an invertible matrix so that is a diagonal matrix.
Two matrices and are similar if there is an invertible matrix so that
.
Similar matrices have the same determinants, same characteristic polynomials, and same eigenvalues. Note that similar matrices do not necessarily have the same eigenvectors corresponding to the same eigenvalues.
An matrix is diagonalizable if and only if has linearly independent eigenvectors.
When an matrix is diagonalizable, then is invertible and is diagonal, where ,,, are linearly independent eigenvectors for .
One use for diagonalization is that once we have diagonalized a matrix we can quickly and easily compute powers of . Diagonalization can also help us understand the actions of matrix transformations.
Determine if each of the following matrices is diagonalizable or not. For diagonalizable matrices, clearly identify a matrix which diagonalizes the matrix, and what the resulting diagonal matrix is.
Suppose a matrix has eigenvalues 2, 3 and 5 and the eigenspace for the eigenvalue 3 has dimension 2. Do we have enough information to determine if is diagonalizable? Explain.
Let and . Find the eigenvalues and eigenvectors of and . Conclude that it is possible for two different matrices and to have exactly the same eigenvectors and corresponding eigenvalues.
A natural question to ask is if there are any conditions under which matrices that have exactly the same eigenvectors and corresponding eigenvalues must be equal. Determine the answer to this question if and are both diagonalizable.
Exercise 1 in Section 18 shows that the determinant of a matrix is the product of its eigenvalues. In this exercise we show that the trace of a diagonalizable matrix is the sum of its eigenvalues.โ35โ First we define the trace of a matrix.
Let be a diagonalizable matrix, and let be the characteristic polynomial of . Let be an invertible matrix such that , where is the diagonal matrix whose diagonal entries are ,,,, the eigenvalues of (note that these eigenvalues may not all be distinct).
The real exponential function satisfies some familiar properties. For example, and for any real numbers and . Does the matrix exponential satisfy the corresponding properties. That is, if and are matrices, must and ? Explain.
Follow the steps indicated to show that if is an matrix and and are any scalars, then . (Although we will not use it, you may assume that the series for converges for any square matrix .)
There is an interesting connection between the determinant of a matrix exponential and the trace of the matrix. Let be a diagonalizable matrix with real entries. Let for some invertible matrix , where is the diagonal matrix with entries ,,, the eigenvalues of .
The first part of this exercise presents an example of a matrix that satisfies its own characteristic equation. Show that if is an diagonalizable matrix with characteristic polynomial , then .โ36โ That is, if , then . (Hint: If for some diagonal matrix , show that . Then determine .)
We return to the Fibonacci sequence where , for ,, and . Since is determined by previous values and , the relation is called a recurrence relation. The recurrence relation is very time consuming to use to compute for large values of . It turns out that there is a fascinating formula that gives the th term of the Fibonacci sequence directly, without using the relation .
So if we can somehow easily find the powers of the matrix , then we can find a convenient formula for . As we have seen, we know how to do this if is diagonalizable
Formula (19.6) is called Binet's formula. It is a very surprising formula in the fact that the expression on the right hand side of (19.6) is an integer for each positive integer . Note that with Binet's formula we can quickly compute for very large values of . For example,
The number , called the golden mean or golden ratio is intimately related to the Fibonacci sequence. Binet's formula provides a fascinating relationship between the Fibonacci numbers and the golden ratio. The golden ratio also occurs often in other areas of mathematics. It was an important number to the ancient Greek mathematicians who felt that the most aesthetically pleasing rectangles had sides in the ratio of .
You might wonder what happens if we use negative integer exponents in Binet's formula. In other words, are there negatively indexed Fibonacci numbers? For any integer , including negative integers, let
There is a specific relationship between and . Find it and verify it.
This result is true for any matrix, but the argument is more complicated.
This result is known as the Cayley-Hamilton Theorem and is one of the fascinating results in linear algebra. This result is true for any square matrix.