Skip to main content

Section 19 Diagonalization

Subsection Application: The Fibonacci Numbers

In 1202 Leonardo of Pisa (better known as Fibonacci) published Liber Abaci (roughly translated as The Book of Calculation), in which he constructed a mathematical model of the growth of a rabbit population. The problem Fibonacci considered is that of determining the number of pairs of rabbits produced in a given time period beginning with an initial pair of rabbits. Fibonacci made the assumptions that each pair of rabbits more than one month old produces a new pair of rabbits each month, and that no rabbits die. (We ignore any issues about that might arise concerning the gender of the offspring.) If we let Fn represent the number of rabbits in month n, Fibonacci produced the model

(19.1)Fn+2=Fn+1+Fn,

for nโ‰ฅ0 where F0=0 and F1=1. The resulting sequence

1,1,2,3,5,8,13,21,โ€ฆ

is a very famous sequence in mathematics and is called the Fibonacci sequence. This sequence is thought to model many natural phenomena such as number of seeds in a sunflower and anything which grows in a spiral form. It is so famous in fact that it has a journal devoted entirely to it. As a note, while Fibonacci's work Liber Abaci introduced this sequence to the western world, it had been described earlier Sanskrit texts going back as early as the sixth century.

By definition, the Fibonacci numbers are calculated by recursion. This is a very ineffective way to determine entries Fn for large n. Later in this section we will derive a fascinating and unexpected formula for the Fibonacci numbers using the idea of diagonalization.

Subsection Introduction

As we have seen when studying Markov processes, each state is dependent on the previous state. If x0 is the initial state and A is the transition matrix, then the nth state is found by Anx0. In these situations, and others, it is valuable to be able to quickly and easily calculate powers of a matrix. We explore a way to do that in this section.

Preview Activity 19.1.

Consider a very simplified weather forecast. Let us assume there are two possible states for the weather: rainy (R) or sunny(S). Let us also assume that the weather patterns are stable enough that we can reasonably predict the weather tomorrow based on the weather today. If is is sunny today, then there is a 70% chance that it will be sunny tomorrow, and if it is rainy today then there is a 40% chance that it will be rainy tomorrow. If x0=[sr] is a state vector that indicates a probability s that it is sunny and probability r that it is rainy on day 0, then

x1=[0.700.400.300.60]x0

tells us the likelihood of it being sunny or rainy on day 1. Let A=[0.700.400.300.60].

(a)

Suppose it is sunny today, that is x0=[10]. Calculate x1=Ax0 and explain how this matrix-vector product tells us the probability that it will be sunny tomorrow.

(b)

Calculate x2=Ax1 and interpret the meaning of each component of the product.

(c)

Explain why x2=A2x0. Then explain in general why xn=Anx0.

(d)

The previous result demonstrates that to determine the long-term probability of a sunny or rainy day, we want to be able to easily calculate powers of the matrix A. Use a computer algebra system (e.g., Maple, Mathematica, Wolfram|Alpha) to calculate the entries of x10, x20, and x30. Based on this data, what do you expect the long term probability of any day being a sunny one?

Subsection Diagonalization

In Preview Activity 19.1 we saw how if we can powers of a matrix we can make predictions about the long-term behavior of some systems. In general, calculating powers of a matrix can be a very difficult thing, but there are times when the process is straightforward.

Activity 19.2 illustrates that calculating powers of square matrices whose only nonzero entries are along the diagonal is rather simple. In general, if

D=[d1100โ‹ฏ000d220โ‹ฏ00โ‹ฎ00โ‹ฑโ‹ฎ000โ‹ฏ0dnn],

then

Dk=[d11k00โ‹ฏ000d22k0โ‹ฏ00โ‹ฎ00โ‹ฑโ‹ฎ000โ‹ฏ0dnnk]

for any positive integer k. Recall that a diagonal matrix is a matrix whose only nonzero elements are along the diagonal (see Definition 8.6). In this section we will see that matrices that are similar to diagonal matrices have some very nice properties, and that diagonal matrices are useful in calculations of powers of matrices.

We can utilize the method of calculating powers of diagonal matrices to also easily calculate powers of other types of matrices.

Activity 19.3.

Let D be any matrix, P an invertible matrix, and let A=Pโˆ’1DP.

(c)

Explain in general why An=Pโˆ’1DnP for positive integers n.

As Activity 19.3 illustrates, to calculate the powers of a matrix of the form Pโˆ’1DP we only need determine the powers of the matrix D. If D is a diagonal matrix, this is especially straightforward.

Subsection Similar Matrices

Similar matrices play an important role in certain calculations. For example, Activity 19.3 showed that if we can write a square matrix A in the form A=Pโˆ’1DP for some invertible matrix P and diagonal matrix D, then finding the powers of A is straightforward. As we will see, the relation A=Pโˆ’1DP will imply that the matrices A and D share many properties.

Definition 19.1.

The nร—n matrix A is similar to the nร—n matrix B if there is an invertible matrix P such that A=Pโˆ’1BP.

Activity 19.4.

Let A=[1120] and B=[220โˆ’1]. Assume that A is similar to B via the matrix P=[2122].

(a)

Calculate det(A) and det(B). What do you notice?

(b)

Find the characteristic polynomials of A and B. What do you notice?

(c)

What can you say about the eigenvalues of A and B? Explain.

(d)

Explain why x=[11] is an eigenvector for A with eigenvalue 2. Is x an eigenvector for B with eigenvalue 2? Why or why not?

Activity 19.4 suggests that similar matrices share some, but not all, properties. Note that if A=Pโˆ’1BP, then B=Qโˆ’1AQ with Q=Pโˆ’1. So if A is similar to B, then B is similar to A. Similarly (no pun intended), since A=Iโˆ’1AI (where I is the identity matrix), then any square matrix is similar to itself. Also, if A=Pโˆ’1BP and B=Mโˆ’1CM, then A=(MP)โˆ’1C(MP). So if A is similar to B and B is similar to C, then A is similar to C. If you have studied relations, these three properties show that similarity is an equivalence relation on the set of all nร—n matrices. This is one reason why similar matrices share many important traits, as the next activity highlights.

Activity 19.5.

Let A and B be similar matrices with A=Pโˆ’1BP.

(a)

Use the multiplicative property of the determinant to explain why det(A)=det(B). So similar matrices have the same determinants.

(b)

Use the fact that Pโˆ’1IP=I to show that Aโˆ’ฮปI is similar to Bโˆ’ฮปI.

(c)

Explain why it follows from (a) and (b) that

det(Aโˆ’ฮปI)=det(Bโˆ’ฮปI).

So similar matrices have the same characteristic polynomial, and the same eigenvalues.

We summarize some properties of similar matrices in the following theorem.

Subsection Similarity and Matrix Transformations

When a matrix is similar to a diagonal matrix, we can gain insight into the action of the corresponding matrix transformation. As an example, consider the matrix transformation T from R2 to R2 defined by T(x)=Ax, where

(19.2)A=[3113].

We are interested in understanding what this matrix transformation does to vectors in R2. First we note that A has eigenvalues ฮป1=2 and ฮป2=4 with corresponding eigenvectors v1=[โˆ’11] and v2=[11]. If we let P=[v1 v2], then you can check that

Pโˆ’1AP=D

and

A=PDPโˆ’1,

where

D=[2004].

Thus,

T(x)=PDPโˆ’1x.

A simple calculation shows that

Pโˆ’1=12[โˆ’1111].

Let us apply T to the unit square whose sides are formed by the vectors e1=[10] and e2=[01] as shown in the first picture in Figure 19.3.

To apply T we first multiply e1 and e2 by Pโˆ’1. This gives us

Pโˆ’1e1=12[โˆ’11] and  Pโˆ’1e2=12[11].

So Pโˆ’1 transforms the standard coordinate system into a coordinate system in which columns of Pโˆ’1 determine the axes, as illustrated in the second picture in Figure 19.3. Applying D to the output scales by 2 in the first component and by 4 in the second component as depicted in the third picture in Figure 19.3. Finally, we apply P to translate back into the standard xy coordinate system as shown in the last picture in Figure 19.3. In this case, we can visualize that when we apply the transformation T to a vector in this system it is just scaled in the Pโˆ’1e1โˆ’Pโˆ’1e2 system by the matrix D. Then the matrix P translates everything back to the standard xy coordinate system.

Figure 19.3. The matrix transformation.

This geometric perspective provides another example of how having a matrix similar to a diagonal matrix informs us about the situation. In what follows we determine the conditions that determine when a matrix is similar to a diagonal matrix.

Subsection Diagonalization in General

In Preview Activity 19.1 and in the matrix transformation example we found that a matrix A was similar to a diagonal matrix whose columns were eigenvectors of A. This will work for a general nร—n matrix A as long as we can find an invertible matrix P whose columns are eigenvectors of A. More specifically, suppose A is an nร—n matrix with n linearly independent eigenvectors v1, v1, โ€ฆ, vn with corresponding eigenvalues ฮป1, ฮป1, โ€ฆ, ฮปn (not necessarily distinct). Let

P=[v1 v2 v3 โ‹ฏ vn].

Then

AP=[Av1 Av2 Av3 โ‹ฏ Avn]=[ฮป1v1 ฮป2v2 ฮป3v3 โ‹ฏ ฮปnvn]=[v1 v2 v3 โ‹ฏ vn][ฮป100โ‹ฏ000ฮป20โ‹ฏ00โ‹ฎโ‹ฎโ‹ฏโ‹ฎโ‹ฎ000โ‹ฏฮปnโˆ’10000โ‹ฏ0ฮปn]=PD.

where

D=[ฮป100โ‹ฏ000ฮป20โ‹ฏ00โ‹ฎโ‹ฎโ‹ฏโ‹ฎโ‹ฎ000โ‹ฏฮปnโˆ’10000โ‹ฏ0ฮปn].

Since the columns of P are linearly independent, we know P is invertible, and so

Pโˆ’1AP=D.

Definition 19.4.

An nร—n matrix A is diagonalizable if there is an invertible nร—n matrix P so that Pโˆ’1AP is a diagonal matrix.

In other words, a matrix A is diagonalizable if A is similar to a diagonal matrix.

IMPORTANT NOTE.

The key notion to the process described above is that in order to diagonalize an nร—n matrix A, we have to find n linearly independent eigenvectors for A. When A is diagonalizable, a matrix P so that Pโˆ’1AP is diagonal is said to diagonalize A.

Activity 19.6.

Find an invertible matrix P that diagonalizes A.

(b)

A=[324202423].

Hint.

The eigenvalues of A are 8 and โˆ’1.

It should be noted that there are square matrices that are not diagonalizable. For example, the matrix A=[1101] has 1 as its only eigenvalue and the dimension of the eigenspace of A corresponding to the eigenvalue is one. Therefore, it will be impossible to find two linearly independent eigenvectors for A.

We showed previously that eigenvectors corresponding to distinct eigenvalue are always linearly independent, so if an nร—n matrix A has n distinct eigenvalues then A is diagonalizable. Activity 19.6 (b) shows that it is possible to diagonalize an nร—n matrix even if the matrix does not have n distinct eigenvalues. In general, we can diagonalize a matrix as long as the dimension of each eigenspace is equal to the multiplicity of the corresponding eigenvalue. In other words, a matrix is diagonalizable if the geometric multiplicity is the same is the algebraic multiplicity for each eigenvalue.

At this point we might ask one final question. We argued that if an nร—n matrix A has n linearly independent eigenvectors, then A is diagonalizable. It is reasonable to wonder if the converse is true โ€” that is, if A is diagonalizable, must A have n linearly independent eigenvectors? The answer is yes, and you are asked to show this in Exercise 6. We summarize the result in the following theorem.

Subsection Examples

What follows are worked examples that use the concepts from this section.

Example 19.6.

Let A=[1โˆ’2103โˆ’10โˆ’22] and B=[120010004]. You should use appropriate technology to calculate determinants, perform any row reductions, or solve any polynomial equations.

(a)

Determine if A is diagonalizable. If diagonalizable, find a matrix P that diagonalizes A.

Solution.

Technology shows that the characteristic polynomial of A is

p(ฮป)=det(Aโˆ’ฮปI3)=(4โˆ’ฮป)(1โˆ’ฮป)2.

The eigenvalues of A are the solutions to the characteristic equation p(ฮป)=0. Thus, the eigenvalues of A are 1 and 4. To find a basis for the eigenspace of A corresponding to the eigenvalue 1, we find the general solution to the homogeneous system (Aโˆ’I3)x=0. Using technology we see that the reduced row echelon form of Aโˆ’I3=[0โˆ’2102โˆ’10โˆ’21] is [01โˆ’12000000]. So if x=[x1x2x3], then the general solution to (Aโˆ’I3)x=0 is

x=[x112x3x3]=x1[100]+x3[0121].

So a basis for the eigenspace of A corresponding to the eigenvalue 1 is

{[1 0 0]T,[0 1 2]T}.

To find a basis for the eigenspace of A corresponding to the eigenvalue 4, we find the general solution to the homogeneous system (Aโˆ’4I3)x=0. Using technology we see that the reduced row echelon form of Aโˆ’4I3=[โˆ’3โˆ’210โˆ’1โˆ’10โˆ’2โˆ’2] is [01โˆ’1011000]. So if x=[x1x2x3], then the general solution to (Aโˆ’4I3)x=0 is

x=[x1 x2 x3]T=[x3 โˆ’x3 x3]T=x3[1 โˆ’1 1]T.

So a basis for the eigenspace of A corresponding to the eigenvalue 4 is

{[1 โˆ’1 0]T}.

Eigenvectors corresponding to different eigenvalues are linearly independent, so the set

{[1 0 0]T,[0 1 2]T,[1 โˆ’1 0]T}

is a basis for R3. Since we can find a basis for R3 consisting of eigenvectors of A, we conclude that A is diagonalizable. Letting

P=[10101โˆ’1021]

gives us

Pโˆ’1AP=[100010004].
(b)

Determine if B is diagonalizable. If diagonalizable, find a matrix Q that diagonalizes B.

Solution.

Technology shows that the characteristic polynomial of B is

p(ฮป)=det(Bโˆ’ฮปI3)=(4โˆ’ฮป)(1โˆ’ฮป)2.

The eigenvalues of B are the solutions to the characteristic equation p(ฮป)=0. Thus, the eigenvalues of B are 1 and 4. To find a basis for the eigenspace of B corresponding to the eigenvalue 1, we find the general solution to the homogeneous system (Bโˆ’I3)x=0. Using technology we see that the reduced row echelon form of Bโˆ’I3=[020000003] is [010001000]. So if x=[x1x2x3], then the general solution to (Bโˆ’I3)x=0 is

x=[x1 x2 x3]T=[x1 0 0]T=x1[1 0 0]T.

So a basis for the eigenspace of B corresponding to the eigenvalue 1 is

{[1 0 0]T}.

To find a basis for the eigenspace of B corresponding to the eigenvalue 4, we find the general solution to the homogeneous system (Bโˆ’4I3)x=0. Using technology we see that the reduced row echelon form of Bโˆ’4I3=[โˆ’3200โˆ’30000] is [100010000]. So if x=[x1x2x3], then the general solution to (Bโˆ’4I3)x=0 is

x=[x1 x2 x3]T=[0 0 x3]T=x3[0 0 1]T.

So a basis for the eigenspace of B corresponding to the eigenvalue 4 is

{[0 0 1]T}.

Since each eigenspace is one-dimensional, we cannot find a basis for R3 consisting of eigenvectors of B. We conclude that B is not diagonalizable.

(c)

Is it possible for two matrices R and S to have the same eigenvalues with the same algebraic multiplicities, but one matrix is diagonalizable and the other is not? Explain.

Solution.

Yes it is possible for two matrices R and S to have the same eigenvalues with the same multiplicities, but one matrix is diagonalizable and the other is not. An example is given by the matrices A and B in this problem.

Example 19.7.

(a)

Is it possible to find diagonalizable matrices A and B such that AB is not diagonalizable? If yes, provide an example. If no, explain why.

Solution.

Let A=[1102] and B=[2โˆ’201]. Since A and B are both diagonal matrices, their eigenvalues are their diagonal entries. With 2 distinct eigenvalues, both A and B are diagonalizable. In this case we have AB=[2โˆ’102], whose only eigenvector is 2. The reduced row echelon form of ABโˆ’2I2 is [0100]. So a basis for the eigenspace of AB is {[1 0]T}. Since there is no basis for R2 consisting of eigenvectors of AB, we conclude that AB is not diagonalizable.

(b)

Is it possible to find diagonalizable matrices A and B such that A+B is not diagonalizable? If yes, provide an example. If no, explain why.

Solution.

Let A=[1302] and B=[2001]. Since A and B are both diagonal matrices, their eigenvalues are their diagonal entries. With 2 distinct eigenvalues, both A and B are diagonalizable. In this case we have A+B=[3303], whose only eigenvector is 3. The reduced row echelon form of (A+B)โˆ’3I2 is [0100]. So a basis for the eigenspace of A+B is {[1 0]T}. Since there is no basis for R2 consisting of eigenvectors of A+B, we conclude that A+B is not diagonalizable.

(c)

Is it possible to find a diagonalizable matrix A such that AT is not diagonalizable? If yes, provide an example. If no, explain why.

Solution.

It is not possible to find a diagonalizable matrix A such that AT is not diagonalizable. To see why, suppose that matrix A is diagonalizable. That is, there exists a matrix P such that Pโˆ’1AP=D, where D is a diagonal matrix. Recall that (Pโˆ’1)T=(PT)โˆ’1. So

D=DT=(Pโˆ’1AP)T=PTAT(Pโˆ’1)T=PTAT(PT)โˆ’1.

Letting A=(PT)โˆ’1, we conclude that

Qโˆ’1ATQ=D.

Therefore, Q diagonalizes AT.

(d)

Is it possible to find an invertible diagonalizable matrix A such that Aโˆ’1 is not diagonalizable? If yes, provide an example. If no, explain why.

Solution.

It is not possible to find an invertible diagonalizable matrix A such that Aโˆ’1 is not diagonalizable. To see why, suppose that matrix A is diagonalizable. That is, there exists a matrix P such that Pโˆ’1AP=D, where D is a diagonal matrix. Thus, A=PDPโˆ’1. Since A is invertible, det(A)โ‰ 0. It follows that det(D)โ‰ 0. So none of the diagonal entries of D can be 0. Thus, D is invertible and Dโˆ’1 is a diagonal matrix. Then

Dโˆ’1=(Pโˆ’1AP)โˆ’1=PAโˆ’1Pโˆ’1

and so Pโˆ’1 diagonalizes Aโˆ’1.

Subsection Summary

  • A matrix D=[dij] is a diagonal matrix if dij=0 whenever iโ‰ j.

  • A matrix A is diagonalizable if there is an invertible matrix P so that Pโˆ’1AP is a diagonal matrix.

  • Two matrices A and B are similar if there is an invertible matrix P so that

    B=Pโˆ’1AP.
  • Similar matrices have the same determinants, same characteristic polynomials, and same eigenvalues. Note that similar matrices do not necessarily have the same eigenvectors corresponding to the same eigenvalues.

  • An nร—n matrix A is diagonalizable if and only if A has n linearly independent eigenvectors.

  • When an nร—n matrix A is diagonalizable, then P=[v1 v2 v3 โ‹ฏ vn] is invertible and Pโˆ’1AP is diagonal, where v1, v2, โ€ฆ, vn are n linearly independent eigenvectors for A.

  • One use for diagonalization is that once we have diagonalized a matrix A we can quickly and easily compute powers of A. Diagonalization can also help us understand the actions of matrix transformations.

Exercises Exercises

1.

Determine if each of the following matrices is diagonalizable or not. For diagonalizable matrices, clearly identify a matrix P which diagonalizes the matrix, and what the resulting diagonal matrix is.

(b)

A=[โˆ’14โˆ’2โˆ’340โˆ’313]

2.

The 3ร—3 matrix A has two eigenvalues ฮป1=2 and ฮป2=3. The vectors [121], [1โˆ’12], and [242] are eigenvectors for ฮป1=2, while the vectors [111],[222] are eigenvectors for ฮป2=3. Find the matrix A.

3.

Find a 2ร—2 non-diagonal matrix A and two different pairs of P and D matrices for which A=PDPโˆ’1.

4.

Find a 2ร—2 non-diagonal matrix A and two different P matrices for which A=PDPโˆ’1 with the same D.

5.

Suppose a 4ร—4 matrix A has eigenvalues 2, 3 and 5 and the eigenspace for the eigenvalue 3 has dimension 2. Do we have enough information to determine if A is diagonalizable? Explain.

6.

Let A be a diagonalizable nร—n matrix. Show that A has n linearly independent eigenvectors.

7.

(a)

Let A=[1101] and B=[1201]. Find the eigenvalues and eigenvectors of A and B. Conclude that it is possible for two different nร—n matrices A and B to have exactly the same eigenvectors and corresponding eigenvalues.

(b)

A natural question to ask is if there are any conditions under which nร—n matrices that have exactly the same eigenvectors and corresponding eigenvalues must be equal. Determine the answer to this question if A and B are both diagonalizable.

8.

(a)

Show that if D and Dโ€ฒ are nร—n diagonal matrices, then DDโ€ฒ=Dโ€ฒD.

(b)

Show that if A and B are nร—n matrices and P is an invertible nร—n matrix such that Pโˆ’1AP=D and Pโˆ’1BP=Dโ€ฒ with D and Dโ€ฒ diagonal matrices, then AB=BA.

9.

Exercise 1 in Section 18 shows that the determinant of a matrix is the product of its eigenvalues. In this exercise we show that the trace of a diagonalizable matrix is the sum of its eigenvalues.โ€‰35โ€‰ First we define the trace of a matrix.

Definition 19.8.

The trace of an nร—n matrix A=[aij] is the sum of the diagonal entries of A. That is,

trace(A)=a11+a22+โ‹ฏ+ann=โˆ‘i=1naii.
(a)

Show that if R=[rij] and S=[sij] are nร—n matrices, then trace(RS)=trace(SR).

(b)

Let A be a diagonalizable nร—n matrix, and let p(ฮป)=det(Aโˆ’ฮปIn) be the characteristic polynomial of A. Let P be an invertible matrix such that Pโˆ’1AP=D, where D is the diagonal matrix whose diagonal entries are ฮป1, ฮป2, โ€ฆ, ฮปn, the eigenvalues of A (note that these eigenvalues may not all be distinct).

(i)

Explain why trace(A)=trace(D).

(ii)

Show that the trace of an nร—n diagonalizable matrix is the sum of the eigenvalues of the matrix.

10.

In this exercise we generalize the result of Exercise 12 in Section 8 to arbitrary diagonalizable matrices.

(a)

Show that if

D=[ฮป100โ‹ฏ00ฮป20โ‹ฏ0โ‹ฎโ‹ฎโ‹ฎโ‹ฑโ‹ฎ000โ‹ฏฮปn],

then

eD=[eฮป100โ‹ฏ00eฮป20โ‹ฏ0โ‹ฎโ‹ฎโ‹ฎโ‹ฑโ‹ฎ000โ‹ฏeฮปn].
(b)

Now suppose that an nร—n matrix A is diagonalizable, with Pโˆ’1AP equal to a diagonal matrix D. Show that eA=PeDPโˆ’1.

11.

Let A=[1100] and let B=[0โˆ’100].

(b)

Calculate eB.

Hint.

Explain why B is not diagonalizable.

(d)

The real exponential function satisfies some familiar properties. For example, exey=eyex and ex+y=exey for any real numbers x and y. Does the matrix exponential satisfy the corresponding properties. That is, if X and Y are nร—n matrices, must eXeY=eYeX and eX+Y=eXeY? Explain.

12.

In Exercise 11 we see that we cannot conclude that eX+Y=eXeY for nร—n matrices X and Y. However, a more limited property is true.

(a)
>

Follow the steps indicated to show that if A is an nร—n matrix and s and t are any scalars, then eAseAt=eA(s+t). (Although we will not use it, you may assume that the series for eA converges for any square matrix A.)

(i)

Use the definition to show that

eAseAt=โˆ‘kโ‰ฅ0โˆ‘mโ‰ฅ0sktmk!m!Ak+m.
(ii)

Relabel and reorder terms with n=k+m to show that

eAseAt=sumnโ‰ฅ01n!Anโˆ‘m=0nn!(nโˆ’m)!m!snโˆ’mtm.
(iii)

Complete the problem using the Binomial Theorem that says

(s+t)n=โˆ‘m=0nn!(nโˆ’m)!m!snโˆ’mtm.
(b)

Use the result of part (a) to show that eA is an invertible matrix for any nร—n matrix A.

13.

There is an interesting connection between the determinant of a matrix exponential and the trace of the matrix. Let A be a diagonalizable nร—n matrix with real entries. Let D=Pโˆ’1AP for some invertible matrix P, where D is the diagonal matrix with entries ฮป1, ฮป2, โ€ฆ, ฮปn the eigenvalues of A.

14.

There is interesting relationship between a matrix and its characteristic equation that we explore in this exercise.

(a)

We first illustrate with an example. Let B=[121โˆ’2].

(i)

Show that ฮป2+ฮปโˆ’4 is the characteristic polynomial for B.

(ii)

Calculate B2. Then compute B2+Bโˆ’4I2. What do you get?

(b)

The first part of this exercise presents an example of a matrix that satisfies its own characteristic equation. Show that if A is an nร—n diagonalizable matrix with characteristic polynomial p(x), then p(A)=0.โ€‰36โ€‰ That is, if p(x)=anxn+anโˆ’1xnโˆ’1+โ‹ฏ+a1x+a0, then p(A)=anAn+anโˆ’1Anโˆ’1+โ‹ฏ+a1A+a0=0. (Hint: If A=PDPโˆ’1 for some diagonal matrix D, show that p(A)=Pp(D)Pโˆ’1. Then determine p(D).)

15.

Label each of the following statements as True or False. Provide justification for your response.

(a) True/False.

If matrix A is diagonalizable, then so is AT.

(b) True/False.

If matrix A is diagonalizable, then A is invertible.

(c) True/False.

If an nร—n matrix A is diagonalizable, then A has n distinct eigenvalues.

(d) True/False.

If matrix A is invertible and diagonalizable, then so is Aโˆ’1.

(e) True/False.

If an nร—n matrix C is diagonalizable, then there exists a basis of Rn consisting of the eigenvectors of C.

(f) True/False.

An nร—n matrix with n distinct eigenvalues is diagonalizable.

(g) True/False.

If A is an nร—n diagonalizable matrix, then there is a unique diagonal matrix such that Pโˆ’1AP=D for some invertible matrix P.

(h) True/False.

If A is an nร—n matrix with eigenvalue ฮป, then the dimension of the eigenspace of A corresponding to the eigenvalue ฮป is nโˆ’rank(Aโˆ’ฮปIn).

(i) True/False.

If ฮป is an eigenvalue of an nร—n matrix A, then eฮป is an eigenvalue of eA. (See Exercise 12 in Section 8 for information on the matrix exponential.)

Subsection Project: Binet's Formula for the Fibonacci Numbers

We return to the Fibonacci sequence Fn where Fn+2=Fn+1+Fn, for nโ‰ฅ0, F0=0, and F1=1. Since Fn+2 is determined by previous values Fn+1 and Fn, the relation Fn+2=Fn+1+Fn is called a recurrence relation. The recurrence relation Fn+2=Fn+1+Fn is very time consuming to use to compute Fn for large values of n. It turns out that there is a fascinating formula that gives the nth term of the Fibonacci sequence directly, without using the relation Fn+2=Fn+1+Fn.

Project Activity 19.7.

The recurrence relationFn+2=Fn+1+Fn gives the equations

(19.3)Fn+1=Fn+Fnโˆ’1(19.4)Fn=Fn.

Let xn=[Fn+1Fn] for nโ‰ฅ0. Explain how the equations (19.3) and (19.3) can be described with the matrix equation

(19.5)xn=Axnโˆ’1,

where A=[1110].

The matrix equation (19.5) shows us how to find the vectors xn using powers of the matrix A:

x1=Ax0x2=Ax1=A(Ax0)=A2x0x3=Ax2=A(A2x0)=A3x0 โ‹ฎโ‹ฎxn=Anx0.

So if we can somehow easily find the powers of the matrix A, then we can find a convenient formula for Fn. As we have seen, we know how to do this if A is diagonalizable

Project Activity 19.8.

Let A=[1110].

(a)

Show that the eigenvalues of A are ฯ†=1+52 and ฯ†โ€•=1โˆ’52.

(b)

Find bases for each eigenspace of A.

Now that we have the eigenvalues and know corresponding eigenvectors for A, we can return to the problem of diagonalizing A.

Project Activity 19.9.

(a)

Why do we know that A is diagonalizable?

(b)

Find a matrix P such that Pโˆ’1AP is a diagonal matrix. What is the diagonal matrix?

Now we can find a formula for the nth Fibonacci number.

Project Activity 19.10.

Since Pโˆ’1AP=D, where D is a diagonal matrix, we also have A=PDPโˆ’1. Recall that when A=PDPโˆ’1, it follows that An=PDnPโˆ’1. Use the equation An=PDnPโˆ’1 to show that

(19.6)Fn=ฯ†nโˆ’ฯ†โ€•n5.
Hint.

We just need to calculate the second component of Anx0.

Formula (19.6) is called Binet's formula. It is a very surprising formula in the fact that the expression on the right hand side of (19.6) is an integer for each positive integer n. Note that with Binet's formula we can quickly compute Fn for very large values of n. For example,

F150=9969216677189303386214405760200.

The number ฯ†=1+52, called the golden mean or golden ratio is intimately related to the Fibonacci sequence. Binet's formula provides a fascinating relationship between the Fibonacci numbers and the golden ratio. The golden ratio also occurs often in other areas of mathematics. It was an important number to the ancient Greek mathematicians who felt that the most aesthetically pleasing rectangles had sides in the ratio of ฯ†:1.

Project Activity 19.11.

You might wonder what happens if we use negative integer exponents in Binet's formula. In other words, are there negatively indexed Fibonacci numbers? For any integer n, including negative integers, let

Fn=ฯ†nโˆ’ฯ†โ€•n5

There is a specific relationship between Fโˆ’n and Fn. Find it and verify it.

This result is true for any matrix, but the argument is more complicated.
This result is known as the Cayley-Hamilton Theorem and is one of the fascinating results in linear algebra. This result is true for any square matrix.