Skip to main content

Section 20 Approximating Eigenvalues and Eigenvectors

Subsection Application: Leslie Matrices and Population Modeling

The Leslie Matrix (also called the Leslie Model) is a powerful model for describing an age distributed growth of a population that is closed to migration. In a Leslie model, it is usually the case that only one gender (most often female) is considered. As an example, we will later consider a population of sheep that is being grown commercially. A natural question that we will address is how we can harvest the population to build a sustainable environment.

When working with populations, the matrices we use are often large. For large matrices, using the characteristic polynomial to calculate eigenvalues is too time and resource consuming to be practical, and we generally cannot find the exact values of the eigenvalues. As a result, approximation techniques are very important. In this section we will explore a method for approximating eigenvalues. The eigenvalues of a Leslie matrix are important because they describe the limiting or steady-state behavior of a population. The matrix and model were introduced by Patrick H. Leslie in β€œOn the Use of Matrices in Certain Population Mathematics”, Leslie, P.H., Biometrika, Volume XXXIII, November 1945, pp. 183-212.

Subsection Introduction

We have used the characteristic polynomial to find the eigenvalues of a matrix, and for each eigenvalue row reduced a corresponding matrix to find the eigenvectors This method is only practical for small matrices β€” for more realistic applications approximation techniques are used. We investigate one such technique in this section, the power method.

Preview Activity 20.1.

Let A=[2653]. Our goal is to find a scalar Ξ» and a nonzero vector v so that Av=Ξ»v.

(a)

If we have no prior knowledge of the eigenvalues and eigenvectors of this matrix, we might just begin with a guess. Let x0=[1 0]T be such a guess for an eigenvector. Calculate Ax0. Is x0 an eigenvector of A? Explain.

(b)

If x0 is not a good approximation to an eigenvector of A, then we need to make a better guess. We have little to work with other than just random guessing, but we can use x1=Ax0 as another guess. We calculated x1 in part 1. Is x1 an eigenvector for A? Explain.

(c)

In parts (a) and (b) you might have noticed that in some sense x1 is closer to being an eigenvector of A than x0 was. So maybe continuing this process will get us closer to an eigenvector of A. In other words, for each positive integer k we define xk as Axkβˆ’1. Before we proceed, however, we should note that as we calculate the vectors x1, x2, x3, …, the entries in the vectors get large very quickly. So it will be useful to scale the entries so that they stay at a reasonable size, which makes it easier to interpret the output. One way to do this is to divide each vector xi by its largest component in absolute value so that all of the entries stay between βˆ’1 and 1. 37  So in our example we have x0=[1 0]T, x1=[2/5 1]T, and x2=[1 25/34]T. Explain why scaling our vectors will not affect our search for an eigenvector.

(d)

Use an appropriate technological tool to find the vectors xk up to k=10. What do you think the limiting vector limkβ†’βˆžxk is? Is this limiting vector an eigenvector of A? If so, what is the corresponding eigenvalue?

Subsection The Power Method

While the examples we present in this text are small in order to highlight the concepts, matrices that appear in real life applications are often enormous. For example, in Google's PageRank algorithm that is used to determine relative rankings of the importance of web pages, matrices of staggering size are used (most entries in the matrices are zero, but the size of the matrices is still huge). Finding eigenvalues of such large matrices through the characteristic polynomial is impractical. In fact, finding the roots of all but the smallest degree characteristic polynomials is a very difficult problem. As a result, using the characteristic polynomial to find eigenvalues and then finding eigenvectors is not very practical in general, and it is often a better option to use a numeric approximation method. We will consider one such method in this section, the power method.

In Preview Activity 20.1, we saw an example of a matrix A=[2653] so that the sequence {xk}, where xk=Axkβˆ’1, converged to a dominant eigenvector of A for an initial guess vector x0=[1 0]T. The vectors xi for i from 1 to 6 (with scaling) are approximately

x1=[0.41]x2=[10.7353]x3=[0.88981]x4=[10.9575]x5=[0.98381]x6=[10.9939].

Numerically we can see that the sequence {xk} approaches the vector [1 1]T, and Figure 20.1 illustrates this geometrically as well.

Figure 20.1. The power method.

This method of successive approximations xk=Axkβˆ’1 is called the power method (since we could write xk as Akx0). Our task now is to show that this method works in general. In the next activity we restrict our argument to the 2Γ—2 case, and then discuss the general case afterwards.

Let A be an arbitrary 2Γ—2 matrix with two linearly independent eigenvectors v1 and v2 and corresponding eigenvalues Ξ»1 and Ξ»2, respectively. We will also assume |Ξ»1|>|Ξ»2|. An eigenvalue whose absolute value is larger than that of any other eigenvalue is called a dominant eigenvalue. Any eigenvector for a dominant eigenvalue is called a dominant eigenvector. Before we show that our method can be used to approximate a dominant eigenvector, we recall that since v1 and v2 are eigenvectors corresponding to distinct eigenvalues, then v1 and v2 are linearly independent. So there exist scalars a1 and a2 such that

x0=a1v1+a2v2.

We have seen that for each positive integer k we can write xn as

(20.1)xk=a1Ξ»1kv1+a2Ξ»2kv2.

With this representation of x0 we can now see why the power method approximates a dominant eigenvector of A.

Activity 20.2.

Assume as above that A is an arbitrary 2Γ—2 matrix with two linearly independent eigenvectors v1 and v2 and corresponding eigenvalues Ξ»1 and Ξ»2, respectively. (We are assuming that we don't know these eigenvectors, but we can assume that they exist.) Assume that Ξ»1 is the dominant eigenvalue for A, x0 is some initial guess to an eigenvector for A, that x0=a1v1+a2v2, and that xk=Axkβˆ’1 for kβ‰₯1.

(a)

We divide both sides of equation (20.1) by Ξ»1k (since Ξ»1 is the dominant eigenvalue, we know that Ξ»1 is not 0) to obtain

(20.2)1Ξ»1kxk=a1v1+a2(Ξ»2Ξ»1)kv2.

Recall that Ξ»1 is the dominant eigenvalue for A. What happens to (Ξ»2Ξ»1)k as kβ†’βˆž? Explain what happens to the right hand side of equation (20.2) as kβ†’βˆž.

(b)

Explain why the previous result tells us that the vectors xk are approaching a vector in the direction of v1 or βˆ’v1 as kβ†’βˆž, assuming a1β‰ 0. (Why do we need a1β‰ 0? What happens if a1=0?)

(c)

What does all of this tell us about the sequence {xk} as kβ†’βˆž?

The power method is straightforward to implement, but it is not without its drawbacks. We began by assuming that we had a basis of eigenvectors of a matrix A. So we are also assuming that A is diagonalizable. We also assumed that A had a dominant eigenvalue Ξ»1. That is, if A is nΓ—n we assume that A has eigenvalues Ξ»1, Ξ»2, …, Ξ»n, not necessarily distinct, with

|Ξ»1|>|Ξ»2|β‰₯|Ξ»3|β‰₯β‹―β‰₯|Ξ»n|

and with vi an eigenvector of A with eigenvalue Ξ»i. We could then write any initial guess x0 in the form

x0=a1v1+a2v2+β‹―+anvn.

The initial guess is also called a seed.

Then

xk=a1Ξ»1kv1+a2Ξ»2kv2+β‹―+anΞ»nkvn

and

(20.3)1Ξ»1kxk=a1v1+a2(Ξ»2Ξ»1)kv2+β‹―+an(Ξ»nΞ»1)kvn.

Notice that we are not actually calculating the vectors xk here β€” this is a theoretical argument and we don't know Ξ»1 and are not performing any scaling like we did in Preview Activity 20.1. We are assuming that Ξ»1 is the dominant eigenvalue of A, though, so for each i the terms (Ξ»iΞ»1)k converge to 0 as k goes to infinity. Thus,

xkβ‰ˆΞ»1ka1v1

for large values of k, which makes the sequence {xk} converge to a vector in the direction of a dominant eigenvector v1 provided a1β‰ 0. So we need to be careful enough to choose a seed that has a nonzero component in the direction of v1. Of course, we generally don't know that our matrix is diagonalizable before we make these calculations, but for many matrices the sequence {xk} will approach a dominant eigenvector.

The power method approximates a dominant eigenvector, and there are ways that we can approximate the dominant eigenvalue. Exercise 8 presents one way β€” by keeping track of the components of the xk that have the largest absolute values, and the next activity shows another.

Activity 20.3.

Let A be an nΓ—n matrix with eigenvalue Ξ» and corresponding eigenvector v.

(a)

Explain why Ξ»=Ξ»(vβ‹…v)vβ‹…v.

(b)

Use the result of part (a) to explain why Ξ»=(Av)β‹…vvβ‹…v.

The result of Activity 20.3 is that, when the vectors in the sequence {xk} approximate a dominant eigenvector of a matrix A, the quotients

(20.4)(Axk)β‹…xkxkβ‹…xk=xkTAxkxkTxk

approximate the dominant eigenvalue of A. The quotients in (20.4) are called Rayleigh quotients.

To summarize, the procedure for applying the power method for approximating a dominant eigenvector and dominant eigenvalue of a matrix A is as follows.

Step 1

Select an arbitrary nonzero vector x0 as an initial guess to a dominant eigenvector.

Step 2

Let x1=Ax0. Let k=1.

Step 3

To avoid having the magnitudes of successive approximations become excessively large, scale this approximation xk. That is, find the entry Ξ±k of xk that is largest in absolute value. Then replace xk by 1Ξ±kxk.

Step 4

Calculate the Rayleigh quotient rk=(Axk)β‹…xkxkβ‹…xk.

Step 5

Let let xk+1=Axk. Increase k by 1 and repeat Steps 3 through 5.

If the sequence {xk} converges to a dominant eigenvector of A, then the sequence {rk} converges to the dominant eigenvalue of A.

The power method can be useful for approximating a dominant eigenvector as long as the successive multiplications by A are fairly simple β€” for example, if many entries of A are zero. 38  The rate of convergence of the sequence {xk} depends on the ratio Ξ»2Ξ»1. If this ratio is close to 1, then it can take many iterations before the power (Ξ»2Ξ»1)k makes the v2 term negligible. There are other methods for approximating eigenvalues and eigenvectors, e.g., the QR factorization, that we will not discuss at this point.

Subsection The Inverse Power Method

The power method only allows us to approximate the dominant eigenvalue and a dominant eigenvector for a matrix A. It is possible to modify this method to approximate other eigenvectors and eigenvalues under certain conditions. We consider an example in the next activity to motivate the general situation.

Activity 20.4.

Let A=[2653] be the matrix from Preview Activity 20.1. Recall that 8 is an eigenvalue for A, and a quick calculation can show that βˆ’3 is the other eigenvalue of A. Consider the matrix B=(Aβˆ’(βˆ’2)I2)βˆ’1=110[βˆ’565βˆ’4].

(a)

Show that 18βˆ’(βˆ’2) and 1βˆ’3βˆ’(βˆ’2) are the eigenvalues of B.

(b)

Recall that v1=[1 1]T is an eigenvector of A corresponding to the eigenvalue 8 and assume that v2=[βˆ’6 5]T is an eigenvector for A corresponding to the eigenvalue βˆ’3. Calculate the products Bv1 and Bv2. How do the products relate to the results of part (a)?

Activity 20.4 provides evidence that we can translate the matrix A having a dominant eigenvalue to a different matrix B with the same eigenvectors as A and with a dominant eigenvalue of our choosing. To see why, let A be an nΓ—n matrix with eigenvalues Ξ»1, Ξ»2, …, Ξ»n, and let Ξ± be any real number distinct from the eigenvalues. Let B=(Aβˆ’Ξ±In)βˆ’1. In our example in Activity 20.4 the numbers

1Ξ»1βˆ’Ξ±, 1Ξ»2βˆ’Ξ±, 1Ξ»3βˆ’Ξ±, β€¦,1Ξ»nβˆ’Ξ±

were the eigenvalues of B, and that if vi is an eigenvector for A corresponding to the eigenvalue Ξ»i, then vi is an eigenvector of B corresponding to the eigenvalue 1Ξ»iβˆ’Ξ±. To see why, let Ξ» be an eigenvalue of an nΓ—n matrix A with corresponding eigenvector v. Let Ξ± be a scalar that is not an eigenvalue of A, and let B=(Aβˆ’Ξ±In)βˆ’1. Now

Av=Ξ»vAvβˆ’Ξ±v=Ξ»vβˆ’Ξ±v(Aβˆ’Ξ±In)v=(Ξ»βˆ’Ξ±)v1Ξ»βˆ’Ξ±v=(Aβˆ’Ξ±In)βˆ’1v.

So 1Ξ»βˆ’Ξ± is an eigenvalue of B with eigenvector v.

Now suppose that A is an nΓ—n matrix with eigenvalues Ξ»1, Ξ»2, …, Ξ»n, and that we want to approximate an eigenvector and corresponding eigenvalue Ξ»i of A. If we can somehow find a value of Ξ± so that |Ξ»iβˆ’Ξ±|<|Ξ»jβˆ’Ξ±| for all jβ‰ i, then |1Ξ»iβˆ’Ξ±|>|1Ξ»jβˆ’Ξ±| for any jβ‰ i. Thus, the matrix B=(Aβˆ’Ξ±In)βˆ’1 has 1Ξ»iβˆ’Ξ± as its dominant eigenvalue and we can use the power method to approximate an eigenvector and the Rayleigh quotient to approximate the eigenvalue 1Ξ»iβˆ’Ξ±, and hence approximate Ξ»i.

Activity 20.5.

Let A=18[7333022βˆ’1015βˆ’2111].

(a)

Apply the power method to the matrix B=(Aβˆ’I3)βˆ’1 with initial vector x0=[1 0 0]T to fill in Table 20.2 (to four decimal places). Use this information to estimate an eigenvalue for A and a corresponding eigenvector.

Table 20.2. Applying the power method to (Aβˆ’I3)βˆ’1
k 10 15 20
xk
xkTAxkxkTxk

(b)

Applying the power method to the matrix B=(Aβˆ’0I3)βˆ’1 with initial vector x0=[1 0 0]T yields the information in Table 20.3 (to four decimal places). Use this information to estimate an eigenvalue for A and a corresponding eigenvector.

Table 20.3. Applying the power method to (Aβˆ’0I3)βˆ’1
k 10 15 20
xk [βˆ’0.33440.66771.0000] [βˆ’0.33330.66661.0000] [βˆ’0.33330.66661.0000]
xkTAxkxkTxk βˆ’1.0014 βˆ’1.0000 βˆ’1.0000

(c)

Applying the power method to the matrix B=(Aβˆ’5I3)βˆ’1 with initial vector x0=[1 0 0]T yields the information in Table 20.4 (to four decimal places). Use this information to estimate an eigenvalue for A and a corresponding eigenvector.

Table 20.4. Applying the power method to (Aβˆ’5I3)βˆ’1
k 10 15 20
xk [0.00001.0000βˆ’1.0000] [0.00001.0000βˆ’1.0000] [0.00001.0000βˆ’1.0000]
xkTAxkxkTxk βˆ’1.0000 βˆ’1.0000 βˆ’1.0000

Subsection Examples

What follows are worked examples that use the concepts from this section.

Example 20.5.

Let A=[123456789].

(a)

Approximate the dominant eigenvalue of A accurate to two decimal places using the power method. Use technology as appropriate.

Solution.

We use technology to calculate the scaled vectors Akx0 for values of k until the components don't change in the second decimal place. We start with the seed x0=[1 1 1]T. For example, to two decimal places we have xk=[0.28 0.64 1.00]T for kβ‰₯20. So we suspect that [0.28 0.64 1.00]T is close to a dominant eigenvector for A. For the dominant eigenvalue, we can calculate the Rayleigh quotients (Axk)β‹…xkxkβ‹…xk until they do not change to two decimal places. For kβ‰₯4, our Rayleigh quotients are all (to two decimal places) equal to 16.12. So we expect that the dominant eigenvalue of A is close to 16.12. Notice that

A[0.28 0.64 1.00]T=[4.56 10.32 16.08]T,

which is not far off from 16.12[0.28 0.64 1.00]T.

(b)

Find the characteristic polynomial p(Ξ») of A. Then find the the root of p(Ξ») farthest from the origin. Compare to the result of part (a). Use technology as appropriate.

Solution.

The characteristic polynomial of A is

p(Ξ»)=βˆ’Ξ»3+15Ξ»2+18Ξ»=βˆ’Ξ»(Ξ»2βˆ’15Ξ»βˆ’18).

The quadratic formula gives the nonzero roots of p(Ξ») as

15Β±152+4(18)2=15Β±3332.

The roots farthest from the origin is approximately 16.12, as was also calculated in part (a).

Example 20.6.

Let A=[210131012].

(a)

Use the power method to approximate the dominant eigenvalue and a corresponding eigenvector (using scaling) accurate to two decimal places. Use x0=[1 1 1]T as the seed.

Solution.

We use technology to calculate the scaled vectors Akx0 for values of k until the components don't change in the second decimal place. For example, to two decimal places we have xk=[0.50 1.00 0.50]T for kβ‰₯4. So we suspect that [12 1 12]T is a dominant eigenvector for A. For the dominant eigenvalue, we can calculate the Rayleigh quotients (Axk)β‹…xkxkβ‹…xk until they do not change to two decimal places. For kβ‰₯2, our Rayleigh quotients are all (to two decimal places) equal to 4. So we expect that the dominant eigenvalue of A is 4. We could also use the fact that

A[12 1 12]T=[2 4 2]T=4[12 1 12]T

to see that [12 1 12]T is a dominant eigenvector for A with eigenvalue 4.

(b)

Determine the exact value of the dominant eigenvalue of A and compare to your result from part (a).

Solution.

Technology shows that the characteristic polynomial of Aβˆ’Ξ»I3 is

p(Ξ»)=βˆ’Ξ»3+7Ξ»2βˆ’14Ξ»+8=βˆ’(Ξ»βˆ’1)(Ξ»βˆ’2)(Ξ»βˆ’4).

We can see from the characteristic polynomial that 4 is the dominant eigenvalue of A.

(c)

Approximate the remaining eigenvalues of A using the inverse power method.

Hint.

Try Ξ±=0.5 and Ξ±=1.8.

Solution.

Applying the power method to B=(Aβˆ’0.5I3)βˆ’1 with seed x0=[1 1 1]T gives xkβ‰ˆ[0.50 1.00 0.50]T for kβ‰₯5, with Rayleigh quotients of 2 (to several decimal places). So 2 is the dominant eigenvalue of B. But 1Ξ»βˆ’0.5 is also the dominant eigenvalue of B, where Ξ» is the corresponding eigenvalue of A. . So to find Ξ», we note that 1Ξ»βˆ’0.5=2 implies that Ξ»=1 is an eigenvalue of A. Now applying the power method to B=(Aβˆ’1.8I3)βˆ’1 with seed x0=[1 1 1]T gives xkβ‰ˆ[1.00 βˆ’1.00 1.00]T for large enough k, with Rayleigh quotients of 5 (to several decimal places). To find the corresponding eigenvalue Ξ» for A, we note that 1Ξ»βˆ’1.8=5, or Ξ»=2 is an eigenvalue of A. Admittedly, this method is very limited. Finding good choices for Ξ± often depends on having some information about the eigenvalues of A. Choosing Ξ± close to an eigenvalue provides the best chance of obtaining that eigenvalue.

Subsection Summary

  • The power method is an iterative method that can be used to approximate the dominant eigenvalue of an nΓ—n matrix A that has n linearly independent eigenvectors and a dominant eigenvalue.

  • To use the power method we start with a seed x0 and then calculate the sequence {xk} of vectors, where xk=Axkβˆ’1. If x0 is chosen well, then the sequence {xk} converges to a dominant eigenvector of A.

  • If A is an nΓ—n matrix with eigenvalues Ξ»1, Ξ»2, …, Ξ»n, to approximate an eigenvector of A corresponding to the eigenvalue Ξ»i, we apply the power method to the matrix B=(Aβˆ’Ξ±In)βˆ’1, where Ξ± is not a eigenvalue of A and |1Ξ»iβˆ’Ξ±|>|1Ξ»jβˆ’Ξ±| for any jβ‰ i.

Exercises Exercises

1.

Let A=[1221]. Let x0=[1 0]T.

(a)

Find the eigenvalues and corresponding eigenvectors for A.

(b)

Use appropriate technology to calculate xk=Akx0 for k up to 10. Compare to a dominant eigenvector for A.

(c)

Use the eigenvectors from part (b) to approximate the dominant eigenvalue for A. Compare to the exact value of the dominant eigenvalue of A.

(d)

Assume that the other eigenvalue for A is close to 0. Apply the inverse power method and compare the results to the remaining eigenvalue and eigenvectors for A.

2.

Let A=[120βˆ’212131]. Use the power method to approximate a dominant eigenvector for A. Use x0=[1 1 1]T as the seed. Then approximate the dominant eigenvalue of A.

3.

Let A=[3βˆ’1βˆ’13]. Use the power method starting with x0=[1 1]T. Explain why the method fails in this case to approximate a dominant eigenvector, and how you could adjust the seed to make the process work.

4.

Let A=[0110].

(a)

Find the eigenvalues and an eigenvector for each eigenvalue.

(b)

Apply the power method with an initial starting vector x0=[0 1]T. What is the resulting sequence?

(c)

Use equation (20.3) to explain the sequence you found in part (b).

5.

Let A=[2653]. Fill in the entries in Table 20.7, where xk is the kth approximation to a dominant eigenvector using the power method, starting with the seed x0=[1 0]T. Compare the results of this table to the eigenvalues of A and limkβ†’βˆžxk+1β‹…xkxkβ‹…xk. What do you notice?

Table 20.7. Values of the Rayleigh quotient
v x0 x1 x2 x3 x4 x5
vTAvvTv            
v x6 x7 x8 x9 x10 x11
vTAvvTv            

6.

Let A=[4βˆ’5215]. The power method will approximate the dominant eigenvalue Ξ»=14. In this exercise we explore what happens if we apply the power method to Aβˆ’1.

(a)

Apply the power method to Aβˆ’1 to approximate the dominant eigenvalue of Aβˆ’1. Use [1 1]T as the seed. How is this eigenvalue related to an eigenvalue of A?

(b)

Explain in general why applying the power method to the inverse of an invertible matrix B might give an approximation to an eigenvalue of B of smallest magnitude. When might this not work?

7.

There are other algebraic methods that do not rely on the determinant of a matrix that can be used to find eigenvalues of a matrix. We examine one such method in this exercise. Let A be any nΓ—n matrix, and let v be any vector in Rn.

(a)

Explain why the vectors

v, Av, A2v, β€¦, Anv

are linearly dependent.

(b)

Let c0, c1, …, cn be scalars, not all 0, so that

c0v+c1Av+c2A2v+β‹―+cnAnv=0.

Explain why there must be a smallest positive integer k so that there are scalars a0, a1, …, ak with akβ‰ 0. such that

a0v+a1Av+a2A2v+β‹―+akAkv=0.
Hint.

Proceed down the list cnβˆ’1, cnβˆ’2, etc., until you reach a weight that is non-zero.

(c)

Let

q(t)=a0+a1t+a2t2+β‹―+aktk.

Then

q(A)=a0+a1A+a2A2+β‹―+akAk

and

q(A)v=(a0+a1A+a2A2+β‹―+akAk)v=a0v+a1Av+a2A2v+β‹―+akAkv=0.

Suppose the polynomial q(t) has a linear factor, say q(t)=(tβˆ’Ξ»)Q(t) for some degree kβˆ’1 polynomial Q(t). Explain why, if Q(A)v is non-zero, Ξ» is an eigenvalue of A with eigenvector Q(A)v.

(d)

This method allows us to find certain eigenvalues and eigenvectors, the roots of the polynomial q(t). Any other eigenvector must lie outside the eigenspaces we have already found, so repeating the process with a vector v not in any of the known eigenspaces will produce different eigenvalues and eigenvectors. Let A=[22βˆ’1222006].

(i)

Find the polynomial q(t). Use v=[1 1 1]T.

(ii)

Find all of the roots of q(t).

(iii)

For each root Ξ» of q(t), find the polynomial Q(t) and use this polynomial to determine an eigenvector of A. Verify your work.

8.

We have seen that the Rayleigh quotients approximate the dominant eigenvalue of a matrix A. As an alternative to using Rayleigh quotients, we can keep track of the scaling factors. Recall that the scaling in the power method can be used to make the magnitudes of the successive approximations smaller and easier to work with. Let A be an nΓ—n matrix and begin with a non-zero seed v0. We now want to keep track of the scaling factors, so let Ξ±0 be the component of v0 with largest absolute value and let x0=1Ξ±0v0. For kβ‰₯0, let vk=Axkβˆ’1, let Ξ±k be the component of vk with largest absolute value and let xk=1Ξ±kvk.

(a)

Let A=[01βˆ’86]. Use x0=[1 1]T as the seed and calculate Ξ±k for k from 1 to 10. Compare to the dominant eigenvalue of A.

(b)

Assume that for large k the vectors xk approach a dominant eigenvector with dominant eigenvalue Ξ». Show now in general that the sequence of scaling factors Ξ±k approaches Ξ».

9.

Let A be an nΓ—n matrix and let Ξ± be a scalar that is not an eigenvalue of A. Suppose that x is an eigenvector of B=(Aβˆ’Ξ±In)βˆ’1 with eigenvalue Ξ². Find an eigenvalue of A in terms of Ξ² and Ξ± with corresponding eigenvector x.

10.

Label each of the following statements as True or False. Provide justification for your response.

(a) True/False.

The largest eigenvalue of a matrix is a dominant eigenvalue.

(b) True/False.

If an nΓ—n matrix A has n linearly independent eigenvectors and a dominant eigenvalue, then the sequence {Akx0} converges to a dominant eigenvector of A for any initial vector x0.

(c) True/False.

If Ξ» is an eigenvalue of an nΓ—n matrix A and Ξ± is not an eigenvalue of A, then Ξ»βˆ’Ξ± is an eigenvalue of Aβˆ’Ξ±In.

(d) True/False.

Every square matrix has a dominant eigenvalue.

Subsection Project: Managing a Sheep Herd

Sheep farming is a significant industry in New Zealand. New Zealand is reported to have the highest density of sheep in the world. Sheep can begin to reproduce after one year, and give birth only once per year. Table 20.8 gives Birth and Survival Rates for Female New Zealand Sheep (from G. Caughley, β€œParameters for Seasonally Breeding Populations,” Ecology, 48, (1967), 834-839). Since sheep hardly ever live past 12 years, we will only consider the population through 12 years.

Table 20.8. New Zealand female sheep data by age group
Age (years) Birth Rate Survival Rate
0-1 0.000 0.845
1-2 0.045 0.975
2-3 0.391 0.965
3-4 0.472 0.950
4-5 0.484 0.926
5-6 0.546 0.895
6-7 0.543 0.850
7-8 0.502 0.786
8-9 0.468 0.691
9-10 0.459 0.561
10-11 0.433 0.370
11-12 0.421 0.000

As sheep reproduce, they add to the 0-1 sheep (lamb) population. The potential to produce offspring is called fecundity (derived from the word fecund which generally refers to reproductive ability) and determines how many lamb are added to the population. Let Fk (the fecundity rate) be the rate at which females in age class k give birth to female offspring. Not all members of a given age group survive to the next age groups, so let sk be the fraction of individuals that survives from age group k to age group k+1. With these ideas in mind, we can create a life cycle chart as in Figure 20.9 that illustrates how the population of sheep changes on a farm (for the sake of space, we illustrate with four age classes).

Figure 20.9. Life cycle with four age classes.

To model the sheep population, we need a few variables. Let n1(0) be the number of sheep in age group 0-1, n2(0) the number in age group 1-2, n3 the number in age group 2-3 and, in general, nk(0) the number of sheep in age group (kβˆ’1)-k at some initial time (time 0), and let

x0=[n1(0) n2(0) n3(0) β‹― n12(0)]T.

We wish to determine the populations in the different groups after one year. Let

x1=[n1(1) n2(1) n3(1) β‹― n12(1)]T,

where n1(1) denotes the number of sheep in age group 0-1, n2(1) the number of sheep in age group 1-2 and, in general, nk(1) the number of tilapia in age group (kβˆ’1)-k after one year.

Project Activity 20.6.

Table 20.8 shows that, on average, each female in age group 1-2 produces 0.045 female offspring in a year. Since there are n2 females in age group 1-2, the lamb population increases by 0.045n2 in a year.

(a)

Continue this analysis to explain why

n1(1)=0.045n2+0.391n3+0.472n4+0.484n5+0.546n6+0.543n7+0.502n8+0.468n9+0.459n10+0.433n11+0.421n12.
(c)

Now explain why

(20.5)x1=Lx0,

where L is the matrix

(20.6)[00.0450.3910.4720.4840.5460.5430.5020.4680.4590.4330.4210.8450000000000000.9750000000000000.9650000000000000.9500000000000000.9260000000000000.8950000000000000.8500000000000000.7860000000000000.6910000000000000.5610000000000000.3700].

Notice that our matrix L has the form

[F1F2F3β‹―Fnβˆ’1Fns100β‹―000s20β‹―0000s3β‹―00β‹±000β‹―snβˆ’10].

Such a matrix is called a Leslie matrix.

Leslie matrices have certain useful properties, and one eigenvalue of a Leslie matrix can tell us a lot about the long-term behavior of the situation being modeled. You can take these properties as fact unless otherwise directed.

  1. A Leslie matrix L has a unique positive eigenvalue Ξ»1 with a corresponding eigenvector v1 whose entries are all positive.

  2. If Ξ»i (i>1) is any other eigenvalue (real or complex) of L, then |Ξ»i|≀λ1. If Ξ»1 is the largest magnitude eigenvalue of a matrix L, we call Ξ»1 a dominant eigenvalue of L.

  3. If any two successive entries in the first row of L are both positive, then |Ξ»i|<Ξ»1 for every i>1. In this case we say that Ξ»1 is a strictly dominant eigenvalue of L. In a Leslie model, this happens when the females in two successive age classes are fertile, which is almost always the case.

  4. If Ξ»1 is a strictly dominant eigenvalue, then xk is approximately a scalar multiple of v1 for large values of k, regardless of the initial state x0. In other words, large state vectors are close to eigenvectors for Ξ»1.

We can use these properties to determine the long-term behavior of the sheep herd.

Project Activity 20.7.

Assume that L is defined by (20.6), and let

xm=[n1(m) n2(m) n3(m) β‹― n12(m)]T,

where n1(m) denotes the number of sheep in age group 0-1, n2(m) the number of sheep in age group 1-2 and, in general, nk(m) the number of sheep in age group (kβˆ’1)-k after k years.

(a)

Assume that x0=[100 100 100 β‹― 100]T. Use appropriate technology to calculate x22, x23, x24, and x25. Round to the nearest whole number. What do you notice about the sheep population? You may use the GeoGebra applet at geogebra.org/m/yqss88xq.

We can use the third and fourth properties of Leslie matrices to better understand the long-term behavior of the sheep population. Since successive entries in the first row of the Leslie matrix in (20.6) are positive, our Leslie matrix has a strictly dominant eigenvalue Ξ»1. Given the dimensions of our Leslie matrix, finding this dominant eigenvalue through algebraic means is not feasible. Use the power method to approximate the dominant eigenvalue Ξ»1 of the Leslie matrix in (20.6) to five decimal places. Explain your process. Then explain how this dominant eigenvalue tells us that, unchecked, the sheep population grows at a rate that is roughly exponential. What is the growth rate of this exponential growth? You may use the GeoGebra applet at geogebra.org/m/yqss88xq.

Project Activity 20.7 indicates that, unchecked, the sheep population will grow without bound, roughly exponentially with ratio equal to the dominant eigenvalue of our Leslie matrix L. Of course, a sheep farmer cannot provide the physical environment or the resources to support an unlimited population of sheep. In addition, most sheep farmers cannot support themselves only by shearing sheep for the wool. Consequently, some harvesting of the sheep population each year for meat and skin is necessary. A sustainable harvesting policy allows for the regular harvesting of some sheep while maintaining the population at a stable level. It is necessary for the farmer to find an optimal harvesting rate to attain this stable population and the following activity leads us through an analysis of how such a harvesting rate can be determined.

Project Activity 20.8.

The Leslie model can be modified to consider harvesting. It is possible to harvest different age groups at different rates, and to harvest only some age groups and not others. In the case of sheep, it might make sense to only harvest from the youngest population since lamb is more desirable than mutton and the lamb population grows the fastest. Assume that this is our harvesting strategy and that we harvest our sheep from only the youngest age group at the start of each year. Let h be the fraction of sheep we harvest from the youngest age group each year after considering growth.

(a)

If we begin with an initial population x0, then the state vector after births and expected deaths is Lx0. Now we harvest. Explain why if we harvest a fraction h from the youngest age group after considering growth, then the state vector after 1 year will be

x1=Lx0βˆ’HLx0,

where

H=[h00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000].
(b)

Our goal is to find a harvesting rate that will lead to a steady state in which the sheep population remains the same each year. In other words, we want to find a value of h, if one exists, that satisfies

(20.7)x=Lxβˆ’HLx.

Show that (20.7) is equivalent to the matrix equation

(20.8)x=(I12βˆ’H)Lx.
(c)

Use appropriate technology to experiment numerically with different values of h to find the value you think gives the best uniform harvest rate. Explain your reasoning. You may use the GeoGebra applet at geogebra.org/m/yqss88xq.

(d)

Now we will use some algebra to find an equation that explicitly gives us the harvest rate in the general setting. This will take a bit of work, but none of it is too difficult. To simplify our work but yet illustrate the overall idea, let us consider the general 4Γ—4 case with arbitrary Leslie matrix

L=[F1F2F3F4s10000s20000s30].

Recall that we want to find a value of h that satisfies (20.8) with H=[h000000000000000]. Let x=[x1 x2 x3 x4]T.

(i)

Calculate the matrix product (I4βˆ’H)L. Explain why this product is again a Leslie matrix and why (I4βˆ’H)L will have a dominant eigenvalue of 1.

(ii)

Now calculate (I4βˆ’H)Lx and set it equal to x. Write down the resulting system of 4 equations that must be satisfied. Be sure that your first equation is

(20.9)x1=(1βˆ’h)F1x1+(1βˆ’h)F2x2+(1βˆ’h)F3x3+(1βˆ’h)F4x4.
(iii)

Equation (20.9) as written depends on the entries of the vector x, but we should be able to arrive at a result that is independent of x. To see how we do this, we assume the population of the youngest group is never 0, so we can divide both sides of (20.9) by x1 to obtain

(20.10)1=(1βˆ’h)F1+(1βˆ’h)F2x2x1+(1βˆ’h)F3x3x1+(1βˆ’h)F4x4x1.

Now we need to write the fractions x2x1, x3x1, and x4x1 so that they do not involve the xi. Use the remaining equations in your system to show that

x2x1=s1x3x1=s1s2x4x1=s1s2s3.
(iv)

Now conclude that the harvesting value h must satisfy the equation

(20.11)1=(1βˆ’h)[F1+F2s1+F3s1s2+F4s1s2s3].

The value R=F1+F2s1+F3s1s2+F4s1s2s3 is called the net reproduction rate of the population and turns out to be the average number of daughters born to a female in her expected lifetime.

(e)

Extend (20.11) to the 12 age group case of the sheep herd. Calculate the value of R for this sheep herd and then find the value of h. Compare this h to the value you obtained through experimentation earlier. Find the fraction of the lambs that should be harvested each year and explain what the stable population state vector x tells us about the sheep population for this harvesting policy.

There are several other ways to scale, but we won't consider them here.
A matrix in which most entries are zero is called a sparse matrix.