Skip to main content

Section 11 The Invertible Matrix Theorem

Subsection Introduction

This section is different than others in this book in that it contains only one long proof of the equivalence of statements that we have already discussed. As such, this is a theoretical section and there is no application connected to it.

The Invertible Matrix Theorem is a theorem that provides many different statements that are equivalent to having a matrix be invertible. To understand the Invertible Matrix Theorem, we need to know what it means for two statements to be equivalent. By equivalent, we mean that if one of the statements is true, then so is the other. We examine this idea in this preview activity.

Preview Activity 11.1.

Let A be an nΓ—n matrix. In this activity we endeavor to understand why the two statements

  1. The matrix A is invertible.

  2. The matrix AT is invertible.

are equivalent. To demonstrate that statements I and II are equivalent, we need to argue that if statement I is true, then so is statement II, and if statement II is true then so is statement I.

(a)

Let's first show that if statement I is true, then so is statement II. So we assume statement I. That is, we assume that A is an invertible matrix. So we know that there is an nΓ—n matrix B such that AB=BA=In, where In is the nΓ—n identity matrix. To demonstrate that statement II must also be true, we need to verify that AT is an invertible matrix.

(ii)

Take the transpose of both sides of the equation AB=In and use the properties of the transpose to write (AB)T in terms of AT and BT.

(iii)

Take the transpose of both sides of the equation BA=In and use the properties of the transpose to write (BA)T in terms of AT and BT.

(iv)

Explain how the previous two parts show that BT is the inverse of AT, so that AT is invertible. So we have shown that if statement I is true, so is statement II. 25 

(b)

Now we prove that if statement II is true, then so is statement I. So we assume statement II. That is, we assume that the matrix AT is invertible. We could do this in the same manner as part (a), or we could be a bit clever. Let's try to be clever.

(ii)

Why can we use the result of part (a) with AT in place of A to conclude that A is invertible? As a consequence, we have demonstrated that A is invertible if AT is invertible. This concludes our argument that statements I and II are equivalent.

Subsection The Invertible Matrix Theorem

We have been introduced to many statements about existence and uniqueness of solutions to systems of linear equations, linear independence of columns of coefficient matrices, onto linear transformations, and many other items. In this section we will analyze these statements in light of how they are related to invertible matrices, with the main goal to understand the Invertible Matrix Theorem.

Recall that an nΓ—n matrix A is invertible if there is an nΓ—n matrix B such that AB=BA=In, where In is the nΓ—n identity matrix. The Invertible Matrix Theorem is an important theorem in that it provides us with a wealth of statements that are all equivalent to the statement that an nΓ—n matrix A is invertible, and connects many of the topics we have been discussing so far this semester into one big picture.

The Invertible Matrix Theorem is a theorem that provides many different statements that are equivalent to a matrix being invertible. As discussed in Preview Activity 11.1, two statements are said to be equivalent if, whenever one of the statements is true, then the other is also true. So to demonstrate, say, statements I and II are equivalent, we need to argue that

  • if statement I is true, then so is statement II, and

  • if statement II is true then so is statement I.

The Invertible Matrix Theorem, however, provides a long list of statements that are equivalent. It would be inefficient to prove, one by one, that each pair of statements is equivalent. (There are (142)=91 such pairs.) Fortunately, there is a shorter method that we can use.

Activity 11.2.

In this activity, we will consider certain parts of the Invertible Matrix Theorem and show that one implies another in a specific order. For all parts in this activity, we assume A is an nΓ—n matrix.

(a)

Consider the following implication: (2)⟹(6): 26  If the equation Ax=0 has only the trivial solution, then the columns of A are linearly independent. This shows that part 2 of the IMT implies part 6 of the IMT. Justify this implication as if it is a T/F problem.

(b)

Justify the following implication: (6)⟹(9): If the columns of A are linearly independent, then the matrix equation Ax=b has exactly one solution for each vector b in Rn.

(c)

Justify the following implication: (9)⟹(4): If the equation Ax=b has exactly one solution for every vector b in Rn, then the columns of A span Rn.

(d)

Justify the following implication: (4)⟹(2): If the columns of A span Rn, then the equation Ax=0 has only the trivial solution.

(e)

Using the above implications you proved, explain why we can conclude the following implication must also be true: (2)⟹(9): If the equation Ax=0 has only the trivial solution, then the matrix equation Ax=b has exactly one solution for each vector b in Rn.

(f)

Using the above implications you proved, explain why any one of the implications (2), (6), (9), and (4) implies any of the others.

Using a similar ordering of circular implications as in Activity 11.2, we can prove the Invertible Matrix Theorem by showing that each statement in the list implies the next statement, and that the last statement implies the first.

Proof of the Invertible Matrix Theorem.

Statement (1) implies Statement (2)

This follows from work done in Activity 11.2.

Statement (2) implies Statement (3)

This was done in Activity 11.2.

Statement (3) implies Statement (4)

Suppose that every column of A is a pivot column. The fact that A is square means that every row of A contains a pivot, and hence the columns of A span Rn.

Statement (4) implies Statement (5)

Since the columns of A span Rn, it must be the case that every row of A contains a pivot. This means that A must be row equivalent to In.

Statement (5) implies Statement (6)

If A is row equivalent to In, there must be a pivot in every column, which means that the columns of A are linearly independent.

Statement (6) implies Statement (7)

If the columns of A are linearly independent, then there is a pivot in every column. Since A is a square matrix, there is a pivot in every row as well. So the columns of A span Rn. Since they are also linearly independent, the columns form a minimal spanning set, which is a basis of Rn.

Statement (7) implies Statement (8)

If the columns form a basis of Rn, then the columns are linearly independent. This means that each column is a pivot column, which also implies Ax=0 has a unique solution and that T is one-to-one.

Statement (8) implies Statement (9)

If T is one-to-one, then A has a pivot in every column. Since A is square, every row of A contains a pivot. Therefore, the system Ax=b is consistent for every b∈Rn and has a unique solution.

Statement (9) implies Statement (10)

If Ax=b has a unique solution for every b, then the transformation T is onto since T(x)=b has a solution for every b.

Statement (10) implies Statement (11)

Assume that T defined by T(x)=Ax is onto. For each i, let ei be the ith column of the nΓ—n identity matrix In. That is, ei is the vector in Rn with 1 in the ith component and 0 everywhere else. Since T is onto, for each i there is a vector ci such that T(ci)=Aci=ei. Let C=[c1 c2 β‹― cn]. Then

AC=A[c1 c2 β‹― cn]=[Ac1 Ac2 β‹― Acn]=[e1 e2 β‹― en]=In.
Statement (11) implies Statement (12)

Assume C is an nΓ—n matrix so that AC=In. First we show that the matrix equation Cx=0 has only the trivial solution. Suppose Cx=0. Then multiplying both sides on the left by A gives us

A(Cx)=A0.

Simplifying this equation using AC=In, we find x=0. Since Cx=0 has only the trivial solution, every column of C must be a pivot column. Since C is an nΓ—n matrix, it follows that every row of C contains a pivot position. Thus, the matrix equation Cx=b is consistent and has a unique solution for every b in Rn. Let vi be the vector in Rn satisfying Cvi=ei for each i between 1 and n and let M=[v1 v2 β‹― vn]. Then CM=In. Now we show that CA=In. Since

AC=In

we can multiply both sides on the left by C to see that

C(AC)=CIn.

Now we multiply both sides on the right by M and obtain

(C(AC))M=CM.

Using the associative property of matrix multiplication and the fact that CM=In shows that

(CA)(CM)=CMCA=In.

Thus, if A and C are nΓ—n matrices and AC=In, then CA=In. So we have proved our implication with D=C

Statement (12) implies Statement (13)

Assume that there is an nΓ—n matrix D so that DA=In. Suppose Ax=0. Then multiplying both sides by A on the left, we find that

D(Ax)=D0(DA)x=0x=0.

So the equation Ax=0 has only the trivial solution and 0 is not an eigenvalue for A.

Statement (13) implies Statement (14)

If 0 is not an eigenvalue of A, then the equation Ax=0 has only the trivial solution. Since statement 2 implies statement 11, there is an nΓ—n matrix C such that AC=In. The proof that statement 11 implies statement 12 shows that CA=In as well. So A is invertible. By taking the transpose of both sides of the equation AAβˆ’1=Aβˆ’1A=In (remembering (AB)T=BTAT) we find

(Aβˆ’1)TAT=AT(Aβˆ’1)T=InT=In.

Therefore, (Aβˆ’1)T is the inverse of AT by definition of the inverse.

Statement (14) implies Statement (1)

Since statement (1) implies statement (14), we proved β€˜If A is invertible, then AT is invertible.’' Using this implication with AT replaced by A, we find that β€˜If AT is invertible, then (AT)T is invertible.’' But (AT)T=A, which proves that statement (14) implies statement (1).

This concludes our proof of the Invertible Matrix Theorem.

Subsection Examples

What follows are worked examples that use the concepts from this section.

Example 11.2.

Let M=[1221010113230100].

(a)

Without doing any calculations, is M invertible? Explain your response.

Solution.

The third column of M is twice the first, so the columns of M are not linearly independent. We conclude that M is not invertible.

(b)

Is the equation Mx=b consistent for every b in R4? Explain.

Solution.

The equation Mx=b is not consistent for every b in R4. If it was, then the columns of M would span R4 and, since there are exactly four columns, the columns of M would be a basis for R4. Thus, M would have to be invertible, which it is not.

(c)

Is the equation Mx=0 consistent? If so, how many solutions does this equation have? Explain.

Solution.

The homogeneous system is always consistent. Since the columns of M are linearly dependent, the equation Mx=0 has infinitely many solutions.

(d)

Is it possible to find a 4Γ—4 matrix P such that PM=I4? Explain.

Solution.

It is not possible to find a 4Γ—4 matrix P such that PM=I4. Otherwise M would have to be invertible.

Example 11.3.

Let M be an nΓ—n matrix whose eigenvalues are all nonzero.

(a)

Let b∈Rn. Is the equation Mx=b consistent? If yes, explain why and find all solutions in terms of M and b. If no, explain why.

Solution.

Since 0 is not an eigenvalue of M, we know that M is invertible. Therefore, the equation Mx=b has the unique solution x=Mβˆ’1b.

(b)

Let S be the matrix transformation defined by S(x)=Mx. Suppose S(a)=S(b) for some vectors a and b in Rn. Must there be any relationship between a and b? If yes, explain the relationship. If no, explain why.

Solution.

The fact that M is invertible implies that S is one-to-one. So if S(a)=S(b), then it must be the case that a=b.

(c)

Let m1, m2, …, mn be the columns of M. In how many ways can we write the zero vector as a linear combination of m1, m2, …, mn? Explain.

Solution.

Because M is invertible, the columns of M are linearly independent. Therefore, there is only the trivial solution to the equation

x1m1+x2m2+β‹―+xnmn=0.

Subsection Summary

  • Two statements are equivalent if, whenever one of the statements is true, then the other must also be true.

  • To efficiently prove that a string of statements are all equivalent, we can prove that each statement in the list implies the next statement, and that the last statement implies the first.

  • The Invertible Matrix Theorem gives us many conditions that are equivalent to an nΓ—n matrix being invertible. This theorem is important because it connects many of the concepts we have been studying.

Exercises Exercises

1.

Consider the matrix A=[12aβˆ’11b11c]. Use the Invertible Matrix Theorem to geometrically describe the vectors [abc] which make A invertible without doing any calculations.

2.

Suppose A is an invertible nΓ—n matrix. Let T be the matrix transformation defined by T(x)=Ax for x in Rn. Show that the matrix transformation S defined by S(x)=Aβˆ’1x is the inverse of the transformation T (i.e., S is the inverse function to T when the transformations are considered as functions).

3.

Label each of the following statements as True or False. Provide justification for your response.

(a) True/False.

If A2 is invertible, then A is invertible.

(b) True/False.

If A and B are square matrices with AB invertible, then A and B are invertible.

(c) True/False.

If the columns of an nΓ—n matrix A span Rn, then the equation Aβˆ’1x=0 has a unique solution.

(d) True/False.

If the columns of A and columns of B form a basis of Rn, then so do the columns of AB.

(e) True/False.

If the columns of a matrix A form a basis of Rn, then so do the rows of A.

(f) True/False.

If the matrix transformation T defined by T(x)=Ax is one-to-one for an nΓ—n matrix A, then the columns of Aβˆ’1 are linearly independent.

(g) True/False.

If the columns of an nΓ—n matrix A span Rn, then so do the rows of A.

(h) True/False.

If there are two nΓ—n matrices A and B such that AB=In, then the matrix transformation defined by T(x)=ATx is one-to-one.

Note that statement I does not have to be true. We are only assuming that IF statement I is true, then statement II must also be true.
The symbol ⟹ is the implication symbol, so (1)⟹(12) is read to mean that statement (1) of the theorem implies statement (12).