By the end of this section, you should be able to give precise and thorough answers to the questions listed below. You may want to keep these questions in mind to focus your thoughts as you complete the section.
What are two ways to describe what it means for a set of vectors in to be linearly independent?
What are two ways to describe what it means for a set of vectors in to be linearly dependent?
If is a set of vectors, what do we mean by a basis for Span?
Given a nonzero set of vectors, how can we find a linearly independent subset of that has the same span as ?
How do we recognize if the columns of a matrix are linearly independent?
How can we use a matrix to determine if a set of vectors is linearly independent?
How can we use a matrix to find a minimal spanning set for a set of vectors in ?
Bézier curves are simple curves that were first developed in 1959 by French mathematician Paul de Casteljau, who was working at the French automaker Citroën. The curves were made public in 1962 by Pierre Bézier who used them in his work designing automobiles at the French car maker Renault. In addition to automobile design, Bézier curves have many other uses. Two of the most common applications of Bézier curves are font design and drawing tools. As an example, the letter “S” in Palatino font is shown using Bézier curves in Figure 6.1. If you've used Adobe Illustrator, Photoshop, Macromedia Freehand, Fontographer, or any other of a number of drawing programs, then you've used Bézier curves. At the end of this section we will see how Bézier curves can be defined using linearly independent vectors and linear combinations of vectors.
In Section 4 we saw how to represent water-benzene-acetic acid chemical solutions with vectors, where the components represent the water, benzene and acid percentages. We then considered a problem of determining if a given chemical solution could be made by mixing other chemical solutions. Suppose we now have three different water-benzene-acetic acid chemical solutions, one with 40% water, 50% benzene and 10% acetic acid, the second with 52% water, 42% benzene and 6% acid, and a third with 46% water, 46% benzene and 8% acid. We represent the first chemical solution with the vector , the second with the vector , and the third with the vector . By combining these three chemical solutions we can make a chemical solution with 43% water, 48% benzene and 9% acid as follows
(See Exercise 5 of Section 4.) Using the third chemical solution (represented by ) uses more information than we actually need to make the desired 43% water, 48% benzene and 9% acid chemical solution because the vector is redundant — all of the material we need to make is contained in and . This is the basic idea behind linear independence — representing information in the most efficient way.
Information is often contained in and conveyed through vectors — especially linear combinations of vectors. In this section we will investigate the concepts of linear dependence and independence of a set of vectors. Our goal is to be able to efficiently determine when a given set of vectors forms a minimal spanning set. A minimal spanning set is a spanning set that contains the smallest number of vectors to obtain all of the vectors in the span. An important aspect of a minimal spanning set is that every vector in the span can be written in one and only one way as a linear combination of the vectors in the minimal spanning set. This will allow us to define the important notion of the dimension of a vector space.
Recall that a linear combination of vectors ,,, in is a sum of scalar multiples of ,,,. That is, a linear combination of the vectors ,,, is a vector of the form
Recall also that the collection of all linear combinations of a set ,,, of vectors in is called the span of the set of vectors. That is, the span Span of the set ,,, of vectors in is the set
For example, a linear combination of vectors and is . All linear combinations of these two vectors can be expressed as the collection of vectors of the form where are scalars. Suppose we want to determine whether is in the span, in other words if is a linear combination of . This means we are looking for such that
By reducing this matrix, we find that there are no solutions of the system, which implies that is not a linear combination of . Note that we can use any names we please for the scalars, say , if we prefer.
In Task 6.1.a we saw that the vector could be written in infinitely many different ways as linear combinations of ,, and . We now ask the question if we really need all of the vectors ,, and to make as a linear combination in a unique way.
Can the vector be written as a linear combination of the vectors and ? If not, why not? If so, in how many ways can be written as a linear combination of and ? Explain.
In Task 6.1.a we saw that could be written in infinitely many different ways as a linear combination of the vectors ,, and . However, the vector could only be written in one way as a linear combination of and . So is in Span and is also in Span. This raises a question — is any vector in Span also in Span. If so, then the vector is redundant in terms of forming the span of ,, and . For the sake of efficiency, we want to recognize and eliminate this redundancy.
In this section we will investigate the concepts of linear independence of a set of vectors. Our goal is to be able to efficiently determine when a given set of vectors forms a minimal spanning set. This will involve the concepts of span and linear independence. Minimal spanning sets are important in that they provide the most efficient way to represent vectors in a space, and will later allow us to define the dimension of a vector space.
In Preview Activity 6.1 we considered the case where we had a set of three vectors, and the vector was in the span of . So the vector did not add anything to the span of . In other words, the set was larger than it needed to be in order to generate the vectors in its span — that is, SpanSpan. However, neither of the vectors in the set could be removed without changing its span. In this case, the set is what we will call a minimal spanning set or basis for Span. There are two important properties that make a basis for Span. The first is that every vector in Span can be written as linear combinations of and (we also use the terminology that the vectors and span Span), and the second is that every vector in Span can be written in exactly one way as a linear combination of and . This second property is the property of linear independence, and it is the property that makes the spanning set minimal.
To make a spanning set minimal, we want to be able to write every vector in the span in a unique way in terms of the spanning vectors. Notice that the zero vector can always be written as a linear combination of any set of vectors using 0 for all of the weights. So to have a minimal or linearly independent spanning set, that is, to have a unique representation for each vector in the span, it will need to be the case that the only way we can write the zero vector as a linear combination of a set of vectors is if all of the weights are 0. This leads us to the definition of a linearly independent set of vectors.
Which of the following sets in or is linearly independent and which is linearly dependent? Why? For the linearly dependent sets, write one of the vectors as a linear combination of the others, if possible.
Task 6.2.a and Task 6.2.c illustrate how we can write one of the vectors in a linearly dependent set as a linear combination of the others. This would allow us to write at least one of the vectors in the span of the set in more than one way as a linear combination of vectors in this set. We prove this result in general in the following theorem.
A set of vectors in is linearly dependent if and only if at least one of the vectors in the set can be written as a linear combination of the remaining vectors in the set.
The statement of Theorem 6.4 is a bi-conditional statement (an if and only if statement). To prove this statement about the set we need to show two things about . One: we must demonstrate that if is a linearly dependent set, then at least one vector in is a linear combination of the other vectors (this is the “only if” part ofthe biconditional statement) and Two: if at least one vector in is a linear combination of the others, then is linearly dependent (this is the “if” part of the biconditional statement). We illustrate the main idea of the proof using a three vector set .
First let us assume that is a linearly dependent set and show that at least one vector in is a linear combination of the other vectors. Since is linearly dependent we can write the zero vector as a linear combination of ,, and with at least one nonzero weight. For example, suppose
(6.1)
Solve Equation (6.1) for the vector to show that can be written as a linear combination of and . Conclude that is a linear combination of the other vectors in the set .
Now we assume that at least one of the vectors in is a linear combination of the others. For example, suppose that
.(6.2)
Use vector algebra to rewrite Equation (6.2) so that is expressed as a linear combination of ,, and such that the weight on is not zero. Conclude that the set is linearly dependent.
Let be a set of vectors in . We will begin by verifying the first statement.
We assume that is a linearly dependent set and show that at least one vector in is a linear combination of the others. Since is linearly dependent, there are scalarc , not all of which are 0, so that
.(6.3)
We don't know which scalar(s) are not zero, but there is at least one. So let us assume that is not zero for some between 1 and . we can then subtract from both sides of Equation (6.3) and divide by to obtain
.
Thus, the vector is a linear combination of ,,,,,,, and at least one of the vectors in is a linear combination of the other vectors in .
To verify the second statement, we assume that at least one of the vectors in can be written as a linear combination of the others and show that is then a linearly dependent set. We don't know which vector(s) in can be written as a linear combination of the others, but there is at least one. Let us suppose that is a linear combination of the others, but there is at least one. Let us suppose that is a linear combination of the vectors ,,,,,,, for some between 1 and . Then there exist scalars ,,,,,, so that
.
It follows that
.
So there are scalars ,,, (with ), not all of which are 0, so that
With a linearly dependent set, at least one of the vectors in the set is a linear combination of the others. With a linearly independent set, this cannot happen — no vector in the set can be written as a linear combination of the others. This result is given in the next theorem. You may be able to see how Theorem 6.4 and Theorem 6.5 are logically equivalent.
A set of vectors in is linearly independent if and only if no vector in the set can be written as a linear combination of the remaining vectors in the set.
As was hinted at in Preview Activity 6.1, an important consequence of a linearly independent set is that every vector in the span of the set can be written in one and only one way as a linear combination of vectors in the set. It is this uniqueness that makes linearly independent sets so useful. We explore this idea in this activity for a linearly independent set of three vectors. Let be a linearly independent set of vectors in for some , and let be a vector in Span. To show that can be written in exactly one way as a linear combination of vectors in , we assume that
and
for some scalars ,,,,, and . We need to demonstrate that ,, and .
use the two different ways of writing as a linear combination of ,, and to come up with a linear combination expressing as a linear combination of these vectors.
Activity 6.4 contains the general ideas to show that any vector in the span of a linearly independent set can be written in one and only one way as a linear combination of the vectors in the set. The weights of such a linear combination provide us a coordinate system for the vectors in terms of the basis. Two familiar concepts of coordinate systems are the Cartesian coordinates and -plane, and -space. We will revisit the coordinate system idea in a later chapter.
In the next theorem we state and prove the general case of any number of linearly independent vectors productin unique representations as linear combinations.
Let be a linearly independent set of vectors in . Any vector in Span can be written in one and only one way as as linear combination of the vectors ,, and .
Let be a linearly independent set of vectors in and let be a vector in Span. By definition, it follows that can be written as a linear combination of the vectors in . It remains for us to show that this representation is unique. So assume that
and (6.4)
for some scalars ,,, and ,,,. Then
.
Subtracting all terms from the right side and using a little vector algebra gives us
.
The fact that is a linearly independent set implies that
,
showing that for every between 1 and . We conclude that the representation of as a linear combination of the linearly independent vectors in is unique.
In this activity we learn how to use a matrix to determine in general if a set of vectors in is linearly independent or dependent. Suppose we have vectors ,,, in . To see if these vectors are linearly independent, we need to find the solutions to the vector equation
.(6.5)
If we let and , then we can write the vector equation (6.5) in matrix form . Let be the reduced row echelon form of .
What can we say about the pivots of in order for to have exactly one solution? Under these conditions, are the vectors ,,, linearly independent or dependent?
What can we say about the rows or columns of in order for to have infinitely many solutions? Under these conditions, are the vectors ,,, linearly independent or dependent?
Use the result of parts (a) and (b) to determine if the vectors ,, and in are linearly independent or dependent. If dependent, write one of the vectors as a linear combination of the others. You may use the fact that the matrix is row equivalent to .
The three concepts — linear independence, span, and minimal spanning set — are different. The important point to note is that minimal spanning set must be both linearly independent and span the space.
To find a minimal spanning set we will often need to find a smallest subset of a given set of vectors that has the same span as the original set of vectors. In this section we determine a method for doing so.
Write the general solution to the homogeneous system , where . Write all linear combinations of ,,, and that are equal to , using weights that only involve and .
Explain how we can conveniently choose the weights in the general solution to to show that the vector is a linear combination of ,, and . What does this tell us about Span and Span?
Explain how we can conveniently choose the weights in the general solution to to show why the vector is a linear combination of and . What does this tell us about Span and Span?
Activity 6.6 illustrates how we can use a matrix to determine a minimal spanning set for a given set of vectors in .
Form the matrix .
Find the reduced row echelon form of . If contains non-pivot columns, say for example that the th column is a non-pivot column, then we can choose the weight corresponding to the th column to be 1 and all weights corresponding to the other non-pivot columns to be 0 to make a linear combination of the columns of that is equal to . This allows us to write as a linear combination of the vectors corresponding to the pivot columns of as we did in the proof of Theorem 6.5. So every vector corresponding to a non-pivot column is in the span of the set of vectors corresponding to the pivot columns. The vectors corresponding to the pivot columns are linearly independent, since the matrix with those columns has every column as a pivot column. Thus, the set of vectors corresponding to the pivot columns of forms a minimal spanning set for .
The set of pivot columns of the reduced row echelon form of will normally not have the same span as the set of columns of , so it is critical that we use columns of ,not in our minimal spanning set.
Activity 6.6 also illustrates a general process by which we can find a minimal spanning set — that is the smallest subset of vectors that has the same span. This process will be useful later when we consider vectors in arbitrary vector spaces. The idea is that if we can write one of the vectors in a set as a linear combination of the remaining vectors, then we can remove that vector from the set and maintain the same span. In other words, begin with the span of a set and follow these steps:
Step 1
If is a linearly independent set, we already have a minimal spanning set.
Step 2
If is not a linearly independent set, then one of the vectors in is a linear combination of the others. Remove that vector from to obtain a new set . It will be the case that SpanSpan.
Step 3
If is a linearly independent set, then is a minimal spanning set. If not, repeat steps 2 and 3 for the set until you arrive at a linearly independent set.
This process is guaranteed to stop as long as the set contains at least one nonzero vector. A verification of the statement in Step 2 that SpanSpan is given in the next theorem.
The result of Theorem 6.7 is that if we have a finite set of vectors in , we can eliminate those vectors that are linear combinations of others until we obtain a smallest set of vectors that still has the same span. As mentioned earlier, we call such a minimal spanning set a basis.
A basis is defined by two characteristics. A basis must span the space in question and a basis must be a linearly independent set. It is the linear independence that makes a basis a minimal spanning set.
So the set , where and spans . Since the columns of are linearly independent, so is the set . Therefore, the set is a basis for . The vector is in the direction of the positive -axis and the vector is in the direction of the positive -axis, so decomposing a vector as a linear combination of and is akin to identifying the vector with the point as we discussed earlier. The set is called the standard basis for .
in . That is, the vector is the vector with a 1 in the th position and 0s everywhere else. Since the matrix has a pivot in each row and column, the set is a basis for . The set is called the standard basis for .
As we will see later, bases 12 are of fundamental importance in linear algebra in that bases will allow us to define the dimension of a vector space and will provide us with coordinate systems.
Is the set linearly independent or dependent. If independent, explain why. If dependent, write one of the vectors in as a linear combination of the other vectors in .
Solution.
We need to know the solutions to the vector equation
.
If the equation has as its only solution (the trivial solution), then the set is linearly independent. Otherwise the set is linearly dependent. To find the solutions to this system, we row reduce the augmented matrix
.
(Note that we really don't need the augmented column of zeros — row operations won't change that column at all. We just need to know that the column of zeros is there.) Technology shows that the reduced row echelon form of this augmented matrix is
.
The reduced row echelon form tells us that the vector equation is consistent, and the fact that there is no pivot in the fourth column shows that the system has a free variable and more than just the trivial solution. We conclude that is linearly dependent. Moreover, the general solution to our vector equation is
is free is free .
Letting and shows that one non-trivial solution to our vector equation is
and .
Thus,
,
or
and we have written one vector in as a linear combination of the other vectors in .
Find a subset of that is a basis for Span. Explain how you know you have a basis.
Solution.
We have seen that the pivot columns in a matrix form a minimal spanning set (or basis) for the span of the columns of . From part (a) we see that the pivot columns in the reduced row echelon form of are the first and second columns. So a basis for the span of the columns of is . Since the elements of are the columns of , we conclude that the set is a subset of that is a basis for Span.
We need to know if the vectors in are linearly independent and span . Technology shows that the reduced row echelon form of
is
.
Since every column of is a pivot column, the set is linearly independent. The fact that there is a pivot in every row of the matrix means that the equation is consistent for every in . Since is a linear combination of the columns of with weights from , tt follows that the columns of span . We conclude that the set is a basis for .
Let , where is a scalar. Are there any values of for which the set is not a basis for ? If so, find all such values of and explain why is not a basis for for those values of .
Solution.
Technology shows that a row echelon form of is
.
The columns of are all pivot columns (hence linearly independent) as long as , and are linearly dependent when . So the only value of for which is not a basis for is .
A set of vectors in is linearly independent if the vector equation
for scalars has only the trivial solution
.
Another way to think about this is that a set of vectors is linearly independent if no vector in the set can be written as a linear combination of the other vectors in the set.
A set of vectors in is linearly dependent if the vector equation
has a nontrivial solution. That is, we can find scalars that are not all 0 so that
.
Another way to think about this is that a set of vectors is linearly dependent if at least one vector in the set can be written as a linear combination of the other vectors in the set.
If is a set of vectors, a subset of is a basis for Span if is a linearly independent set and SpanSpan.
Given a nonzero set of vectors, we can remove vectors from that are linear combinations of remaining vectors in to obtain a linearly independent subset of that has the same span as .
The columns of a matrix are linearly independent if the equation has only the trivial solution .
The set is linearly independent if and only if every column of the matrix , is a pivot column.
If , then the vectors in the pivot columns of form a minimal spanning set for Span.
Is the set consisting of these vectors linearly independent? If so, explain why. If not, make a single change in one of the vectors so that the set is linearly independent.
In a lab, there are three different water-benzene-acetic acid solutions: The first one with 36% water, 50% benzene and 14% acetic acid; the second one with 44% water, 46% benzene and 10% acetic acid; and the last one with 38% water, 49% benzene and 13% acid. Since the lab needs space, the lab coordinator wants to determine whether all solutions are needed, or if it is possible to create one of the solutions using the other two. Can you help the lab coordinator?
Give an example of vectors such that a minimal spanning set for Span is equal to that of Span; and an example of three vectors such that a minimal spanning set for Span is equal to that of Span.
Bézier curves can be created as linear combinations of vectors. In this section we will investigate how cubic Bézier curves (the ones used for fonts) can be realized through linear and quadratic Bézier curves. We begin with linear Bézier curves.
Start with two vectors and . Linear Bézier curves are linear combinations
of the vectors and for scalars between 0 and 1. (You can visualize these linear combinations using the GeoGebra file Linear Bezier at geogebra.org/m/HvrPhh86. With this file you can draw the vectors for varying values of . You can move the points and in the GeoGebra file, and the slider controls the values of . The point identified with is traced as is changed.) For this activity, we will see what the curve corresponds to by evaluating certain points on the curve in a specific example. Let and .
For each value of , the vector is a linear combination of the vectors and . Note that when , we have and when we have , and for Project Activity 6.8 shows that the vectors trace out the line segment from to . The span of the vectors and for is a linear Bézier curve. Once we have a construction like this, it is natural in mathematics to extend it and see what happens. We do that in the next activity to construct quadratic Bézier curves.
Let ,, and be vectors in the plane. We can then let
and
be the linear Bézier curves as defined in Project Activity 6.8. Since and are vectors, we can define as
.
(You can visualize these linear combinations using the GeoGebra file Quadratic Bezier at geogebra.org/m/VWCZZBXz. With this file you can draw the vectors for varying values of . You can move the points ,, and in the GeoGebra file, and the slider controls the values of . The point identified with is traced as is changed.) In this activity we investigate how the vectors change as changes. For the remainder of this activity, let ,, and .
The span of the vectors and , or the set of points traced out by the vectors for , is a quadratic Bézier curve. To understand why this curve is called quadratic, we examine the situation in a general context in the following activity.
Let ,, and be arbitrary vectors in the plane. Write as a linear combination of ,, and . That is, write in the form for some scalars (that may depend on ) ,, and . Explain why the result leads us to call these vectors quadratic Bézier curves.
Notice that if any one of the lies on the line determined by the other two vectors, then the quadratic Bézier curve is just a line segment. So to obtain something non-linear we need to choose our vectors so that that doesn't happen.
Quadratic Bézier curves are limited, because their graphs are parabolas. For applications we need higher order Bézier curves. In the next activity we consider cubic Bézier curves.
Start with four vectors ,,, — the points defined by these vectors are called control points for the curve. As with the linear and quadratic Bézier curves, we let
and .
Then let
and .
We take this one step further to generate the cubic Bézier curves by letting
.
(You can visualize these linear combinations using the GeoGebra file Cubic Bezier at geogebra.org/m/EDAhudy9. With this file you can draw the vectors for varying values of . You can move the points ,,, and in the GeoGebra file, and the slider controls the values of . The point identified with is traced as is changed.) In this activity we investigate how the vectors change as changes. For the remainder of this activity, let ,,, and .
The span of the vectors and , or the set of points traced out by the vectors for , is a cubic Bézier curve. To understand why this curve is called cubic, we examine the situation in a general context in the following activity.
Let ,,, and be arbitrary vectors in the plane. Write as a linear combination of ,,, and . That is, write in the form for some scalars (that may depend on ) ,,, and . Explain why the result leads us to call these vectors cubic Bézier curves.
Just as with the quadratic case, we need certain subsets of the set of control vectors to be linearly independent so that the cubic Bézier curve does not degenerate to a quadratic or linear Bézier curve.
More complicated and realistic shapes can be represented by piecing together two or more Bézier curves as illustrated with the letter “S” in Figure 6.1. Suppose we have two cubic Bézier curves, the first with control points ,,, and and the second with control points ,,, and . You may have noticed that lies on the tangent line to the first Bézier curve at and that lies on the tangent line to the first Bézier curve at . (Play around with the program Cubic Bezier to convince yourself of these statements. This can be proved in a straightforward manner using vector calculus.) So if we want to make a smooth curve from these two Bézier curves, the curves will need to join together smoothly at and . This will force and the tangents at will have to match. This implies that ,, and all have to lie on this common tangent line. Keeping this idea in mind, use the GeoGebra file Cubic Bezier Pair at geogebra.org/m/UwxQ6RPk to find control points for the pair of Bézier curves that create your own letter S.