By the end of this section, you should be able to give precise and thorough answers to the questions listed below. You may want to keep these questions in mind to focus your thoughts as you complete the section.
What is an orthogonal set in ? What is one useful fact about orthogonal subsets of ?
What is an orthogonal basis for a subspace of ? What is an orthonormal basis?
How does orthogonality help us find the weights to write a vector as a linear combination of vectors in an orthogonal basis?
What is an orthogonal matrix and why are orthogonal matrices useful?
An aircraft in flight, like a plane or the space shuttle, can perform three independent rotations: roll, pitch, and yaw. Roll is a rotation about the axis through the nose and tail of the aircraft, pitch is rotation moving the nose of the aircraft up or down through the axis from wingtip to wingtip, and yaw is the rotation when the nose of the aircraft turns left or right about the axis though the plane from top to bottom. These rotations take place in -space and the axes of the rotations change as the aircraft travels through space. To understand how aircraft maneuver, it is important to know about general rotations in space. These are more complicated than rotations in -space, and, as we will see later in this section, involve orthogonal sets.
If is a basis for a subspace of , we know that any vector in can be written uniquely as a linear combination of the vectors in . In the past, the way we have found the coordinates of with respect to , i.e. the weights needed to write a vector as a linear combination of the elements in , has been to row reduce the matrix to solve the corresponding system. This can be a cumbersome process, especially if we need to do it many times. This process also forces us to determine all of the weights at once. For certain types of bases, namely the orthogonal and orthonormal bases, there is a much easier way to find individual weights for this linear combination.
Recall that two nonzero vectors and in are orthogonal if . We can extend this idea to an entire set. For example, the standard basis for has the property that any two distinct vectors in are orthogonal to each other. The basis vectors in make a very nice coordinate system for , where the basis vectors provide the directions for the coordinate axes. We could rotate this standard basis, or multiply any of the vectors in the basis by a nonzero constant, and retain a basis in which all distinct vectors are orthogonal to each other (e.g., ). We define this idea of having all vectors be orthogonal to each other for sets, and then for bases.
Suppose has coordinates with respect to the basis , i.e.
.
Substitute this expression for in and use the orthogonality property of the basis to show that . Then determine and similarly. Finally, calculate the values of ,, and if .
We defined orthogonal sets in and bases of subspaces of in Definitions 24.1 and Definition 24.2. We saw that the standard basis in is an orthogonal set and an orthogonal basis of β there are many other examples as well.
Assuming that the reduced row echelon form of the matrix is , explain why it is not possible to find a nonzero vector so that is an orthogonal subset of .
The example from Activity 24.2 suggests that we can have three orthogonal nonzero vectors in , but no more. Orthogonal vectors are, in a sense, as far apart as they can be. So we might expect that there is no linear relationship between orthogonal vectors. The following theorem makes this clear.
Let be a set of nonzero orthogonal vectors in . To show that ,,, are linearly independent, assume that
(24.1)
for some scalars ,,,. We will show that for each from 1 to . Since the vectors in are orthogonal to each other, we know that whenever . Fix an index between 1 and . We evaluate the dot product of both sides of (24.1) with and simplify using the dot product properties:
.(24.2)
Now all of the dot products on the left are 0 except for , so (24.2) becomes
.
We assumed that and since , we conclude that . We chose arbitrarily, so we have shown that for each between 1 and . Therefore, the only solution to equation (24.1) is the trivial solution with and the set is linearly independent.
Orthogonality is a useful and important property for a basis to have. In Preview Activity 24.1 we saw that if a vector in the span of an orthogonal basis could be written as a linear combination of the basis vectors as , then . If we continued that same argument we could show that
We can apply this idea in general to see how the orthogonality of an orthogonal basis allows us to quickly and easily determine the weights to write a given vector as a linear combination of orthogonal basis vectors. To see why, let be an orthogonal basis for a subspace of and let be any vector in . We know that
The decomposition (24.3) is even simpler if for each , that is, if is a unit vector for each . In this case, the denominators are all 1 and we don't even need to consider them. We have a familiar example of such a basis for , namely the standard basis .
so the condition implies that the vector has norm 1. An orthogonal basis with this additional condition is a very nice basis and is given a special name.
In other words, an orthonormal basis is an orthogonal basis in which every basis vector is a unit vector. A good question to ask here is how we can construct an orthonormal basis from an orthogonal basis.
We have seen in the diagonalization process that we diagonalize a matrix with a matrix whose columns are linearly independent eigenvectors of . In general, calculating the inverse of the matrix whose columns are eigenvectors of in the diagonalization process can be time consuming, but if the columns form an orthonormal set, then the calculation is very straightforward.
With the result of part (a) in mind, what is the matrix product ? What does this tell us about the relationship between and ? Use technology to calculate and confirm your answer.
The result of Activity 24.5 is that if the columns of a square matrix form an orthonormal set, then . This makes calculating very easy. Note, however, that this only works if the columns of form an orthonormal basis for Col . You should also note that if is an matrix satisfying , then the columns of must form an orthonormal set. Matrices like this appear quite often and are given a special name.
Use the fact that to find . Explain how this shows that the inverse of a rotation matrix by an angle is just another rotation matrix but by the angle .
so . This means that preserves length. Such a transformation is called an isometry and it is convenient to work with functions that don't expand or contract things. Moreover, if and are nonzero vectors, then
Using the dot product formula, we see that if and that for each . Since orthogonal vectors are linearly independent, the set is a linearly independent set with vectors in a -dimensional space. It follows that is an orthonormal basis for .
Show that the result of part (c) is true in general. That is, if is an orthonormal basis for , and if , then
.
Solution.
Let be an orthonormal basis for , and suppose that . Then
.(24.4)
Since is an orthonormal basis for , it follows that if and . Expanding the dot product in (24.4), the only terms that won't be zero are the ones that involve . This leaves us with
Use the result of problem Exercise 3 above and that to show that for a subspace of . (See Exercise 13 in Section 12 for the definition of the sum of subspaces.)
Let be an matrix. We showed that if is an orthogonal matrix, then for any vectors and in . Now we ask if the converse of this statement is true. That is, determine the validity of the following statement: if for any vectors and in , then is an orthogonal matrix? Verify your answer.
In this exercise we examine reflection matrices. In the following exercise we will show that the reflection and rotation matrices are the only orthogonal matrices. We will determine how to represent the reflection across a line through the origin in as a matrix transformation. The setup is as follows. Let be the line through the origin in that makes an angle with the positive -axis as illustrated in Figure 24.10.
Let be an arbitrary vector in as represented in Figure 24.10. Determine the components of the vectors proj and proj. Reproduce Figure 24.10 and draw the vectors proj and proj in your figure.
In this exercise we will show that the only orthogonal matrices are the rotation matrices and the reflection matrices (see Exercise 6). Throughout this exercise let ,,, and be real numbers such that is an orthogonal matrix. Let and be the columns of .
Explain why the terminal point of in standard position lies on the unit circle. Then explain why there is an angle such that and . What angle, specifically, is ? Draw a picture to illustrate.
A similar argument to (b) shows that there is an angle such that . Given that is an orthogonal matrix, how must be related to ? Use this result to find the two possibilities for as a vector in terms of and . (You will likely want to look up some trigonometric identities for this part of the problem.)
By considering the two possibilities from part (c), show that is either a rotation matrix or a reflection matrix. Conclude that the only orthogonal matrices are the reflection and rotation matrices.
Recall that a counterclockwise rotation of -space around the origin by an angle is accomplished by left multiplication by the matrix . Notice that the columns of this rotation matrix are orthonormal, so this rotation matrix is an orthogonal matrix. As the next activity shows, rotation matrices in 3D are also orthogonal matrices.
Let be a rotation matrix in 3D. A rotation does not change lengths of vectors, nor does it change angles between vectors. Let ,, and be the standard unit vectors in .
By Project Activity 24.7 we know that the determinant of any rotation matrix is either or . Having a determinant of preserves orientation, and we will identify these rotations as being counterclockwise, and we will identify the others with determinant of as being clockwise. We will set the convention that a rotation is always measured counterclockwise (as we did in ), and so every rotation matrix will have determinant .
Returning to the counterclockwise rotation of -space around the origin by an angle determined by left multiplication by the matrix , we can think of this rotation in -space as the rotation that keeps points in the plane in the plane, but rotates these points counterclockwise around the axis. In other words, in the standard coordinate system, with standard basis ,,, our rotation matrix has the property that . Now is the third column of , so the third column of is . Similarly, is the first column of and is the second column of . Since is a counterclockwise rotation of the plane space around the origin by an angle it follows that this rotation is given by the matrix
In this notation in (24.5), the subscript gives the direction of the line fixed by the rotation and the angle provides the counterclockwise rotation in the plane perpendicular to this vector. This vector is called a normal vector for the rotation. Note also that the columns of form an orthogonal set such that each column vector has norm .
This idea describes a general rotation matrix in 3D by specifying a normal vector and an angle . For example, with roll, a normal vector points from the tail of the aircraft to its tip. It is our goal to understand how we can determine an arbitrary rotation matrix of the form . We can accomplish this by using the rotation around the axis and change of basis matrices to find rotation matrices around other axes. Let be the standard basis for
Define a new ordered basis so that our axis of rotation is the third vector. So in this case the third vector in will be . The other two vectors need to make an orthonormal set. So we have plenty of choices. For example, we could set . Find the change of basis matrix from to .
To find our rotation matrix around the axis, we can first change basis from to , then perform a rotation around the new axis using (24.5), then changing basis back from to . In other words,
The difference between the two options is that, in the first we have while in the second. Using will give clockwise rotations while gives counterclockwise rotations (this is the difference between a left hand system and a right hand system). So it is important to ensure that our change of basis matrix is one with determinant .
In this activity we find the rotation around the axis given by the line . This line is in the direction of the vector . So we start with making a unit vector in the direction of as the third vector in an ordered basis . The other two vectors need to make an orthonormal set with .
Let be as in the previous part. Now the trick is to find a third unit vector so that is an orthonormal set. This can be done with the cross product. If and , then the cross product of and is the vector
.
(You can check that is an orthogonal set that gives the correct determinant for the change of basis matrix.) Use the cross product to find a unit vector so that is an orthonormal set.
In the next activity we summarize the general process to find a 3D rotation matrix for any normal vector . There is a GeoGebra applet at geogebra.org/m/n9gbjhfx that allows you to visualize rotation matrices in 3D.
Let be a normal vector (nonzero) for our rotation. We need to create an orthonormal basis where is a unit vector in the direction of so that the change of basis matrix has determinant .
Once we have a normal vector and a vector orthogonal to , the vector gives us an orthogonal set . We then normalize each vector to create our orthonormal basis . Use this process to find the matrix that produces a counterclockwise rotation around the normal vector .
It isn't clear why such matrices are called orthogonal since the columns are actually orthonormal, but that is the standard terminology in mathematics.