Section 17.3 Concepts
Statement 3 of Proposition 16.5.5 guarantees that every vector space has a spanning set. To prove this statement, we verified that every vector space is trivially generated by itself, i.e.
But this doesn’t give us a useful description of the vector space it basically just says “to build all the vectors in out of some vectors in use all the vectors in ” The point of a spanning set is to give you a set of building-block vectors that can be used to construct every other vector in the space through linear combinations. To make an analogy to chemistry, the vectors in a spanning set act like atoms, and linear combinations are then like molecules built out of different combinations of different quantities of those atoms. So we would like our set of building blocks to be as simple as possible — that is, we would like to get down to some sort of optimal description of a vector space in terms of a spanning set. Linear dependence and independence are precisely the concepts we will need in order to judge whether we have such an optimal spanning set.
Subsection 17.3.1 Reducing spanning sets
Suppose we have a spanning set for a space so that This equality of spaces says that every vector in the space can somehow be expressed as a linear combination of vectors in
Suppose further that one of the vectors in can be expressed as a linear combination of some of the others, say
Here, each of the vectors and the vectors are in making this an expression for as a linear combination of vectors in but not including .
If isn’t needed to express vectors in as linear combinations of vectors in then we should have where is the set of all vectors in except That is, we can discard from the spanning set for and still have a spanning set.
This pattern will help us reduce down to some sort of optimal spanning set: we can keep removing these redundant spanning vectors that are linear combinations of other spanning vectors until there are none left.
Subsection 17.3.2 Linear dependence and independence
A set of vectors is called linearly dependent precisely when it is nonoptimal as a spanning set in the way described above: when one of the vectors in the set is a linear combination of others in the set. However, it is usually not obvious that some vector is redundant in this way — checking each vector in turn is tedious, and also would not be a very convenient way to reason with the concept of linear dependence in the abstract.
is equivalent to having a nontrivial linear combination of these vectors equal to the zero vector:
And vice versa, since if we have a nontrivial linear combination of these vectors that results in the zero vector, say
an expression for as a linear combination of the others.
Again, the advantage of checking for linear combinations equal to the zero vector is that in a collection of vectors we usually don’t know ahead of time which one will be the odd one out (that is, which one will play the role of as above). In a nontrivial linear combination equalling we can take as the redundant vector any of the vectors whose coefficient is nonzero (which is required to perform the step of dividing by in the algebra isolating above).
Now, we can always take the trivial linear combination, where all coefficients are to get a result of But if this is the only linear combination of a set of vectors by which we can get as a result, then none of the vectors can act as the redundant as above, because an expression like (✶✶) always leads to an equality like (✶✶✶), involving a nontrivial linear combination.
This logic leads to the Test for Linear Dependence/Independence.
Procedure 17.3.1. Test for Linear Dependence/Independence.
To test whether vectors are linearly dependent/independent, solve the homogeneous vector equation
in the (scalar) variables
If this vector equation has a nontrivial solution for these coefficient variables, then the vectors are linearly dependent.
Otherwise, if this vector equation has only the trivial solution then the vectors are linearly independent.
Subsection 17.3.3 Linear dependence and independence of just one or two vectors
For a pair of vectors to be linearly dependent, one must be a linear combination of the other. But a linear combination of one vector is just a scalar multiple, and so a pair of vectors is linearly dependent if one is a scalar multiple of the other. If both vectors are nonzero, that scalar must also be nonzero and so could be shifted to the other side as its reciprocal:
So nonzero vectors are linearly dependent if and only if each is a scalar multiple of the other. In we would have called such vectors parallel.
What about a set containing a single vector? Our motivation for creating the concept of linear dependence/independence was to measure whether a spanning set contained redundent information or whether it was somehow “optimal.” A spanning set consisting of a single nonzero vector cannot be reduced to a smaller spanning set, so it is already optimal and we should refer to that spanning set as linearly independent. This coincides with the result of the test for linear dependence/independence for a single vector if is nonzero, then there are no nontrivial solutions to the vector equation
But what about the zero vector by itself? Scalar multiples of remain so is the trivial vector space consisting of just the zero vector. Is an optimal spanning set for the trivial space, or can it be reduced further? By convention, we also consider to be the trivial vector space (where represents a set containing no vectors, called the empty set), because we always want the process to return a subspace of the vector space in which we’re working. So the spanning set can be reduced to the empty set, and still span the same space. This line of reasoning leads us to consider the set of vectors containing only the zero vector to be linearly dependent.
Subsection 17.3.4 Linear dependence and independence in
Independent directions.
In Subsection 16.3.3, we discussed how a nonzero vector in spans a line through the origin. If a second vector is linearly dependent with the vector spanning the line, then as discussed in the previous subsection (Subsection 17.3.3), that second vector must be parallel with the first, hence parallel with the line. To get something linearly independent, we need to branch off in a new direction from the line.
Diagram of linear independence in the case of two vectors.
In Subsection 16.3.3, we also discussed how a pair of nonzero, nonparallel vectors in span a plane through the origin. If a third vector is linearly dependent with those two spanning vectors, it is somehow a linear combination of them and so lies in that plane. To get something linearly independent, we need to branch off in a new direction from that plane.
Diagram of linear independence in the case of three vectors.
This idea that we can enlarge an independent set by including a new vector in a new direction works in abstract vector spaces as well, as we will see in Proposition 17.5.6 in Subsection 17.5.2.
Maximum number of linearly independent vectors.
In Discovery 17.5, we considered the possible sizes of linearly independent sets in and We can certainly have two linearly independent vectors in since clearly the standard basis vectors and form a linearly independent set in Could we have three? Two linearly independent vectors would have to be nonparallel, and so they would have to span a plane — i.e. they would have to span the plane. Geometrically, a third linearly independent vector would have to branch off in a “new direction,” as in our discussion above. But in there is no new third direction in which to head — we can’t have a vector breaking up out of the plane “into the third dimension,” because there is no third dimension available in Algebraically, if we had three vectors in and attempted the test for dependence/independence, we would start by setting up the vector equation
Combining the linear combination on the left back into one vector, and comparing coordinates on either side, we would obtain linear system
in the unknown coefficients Because there are only two equations, the reduced form for the coefficient matrix for this system can have no more than two leading ones, so it requires at least one parameter to solve, which means there are nontrivial solutions. So three vectors in can never by linearly independent.
We come to a similar conclusion in using both geometric and algebraic reasoning — three independent vectors in is certainly possible (for example, the standard basis vectors), but a set of four vectors in can never be linearly independent. Geometrically, once you have three independent vectors pointing in three “independent directions,” there is no new direction in for a fourth independent vector to take. Algebraically, we could set up the test for independence with four vectors in and it would lead to a homogeneous system of three equations (one for each coordinate) in four variables (one unknown coefficient for each vector). Since the system would have more variables than equations, it would require parameters to solve, leading to nontrivial solutions.
And the pattern continues in higher dimensions — a collection of more than four vectors in must be linearly dependent, a collection of more than five vectors in must be linearly dependent, and so on. In fact, this pattern can also be found in abstract vectors spaces — see Lemma 17.5.7 in Subsection 17.5.2. And this pattern will help us transplant the idea of dimension from to abstract spaces.