Section 12.3 Concepts
In this section.
Subsection 12.3.1 Vectors
A directed line segment (or arrow) could be thought of dynamically as describing a change in position, from the initial point to the terminal point. A two-dimensional vector in the plane or a three-dimensional vector in space captures just the change part of “change in position,” leaving the position part (that is, the initial and terminal points) unspecified. For example, in the plane, the instructions “move two units right and three units down” describe a way to change positions, but don't actually specify from where or to where the change in position is occurring. So a vector corresponds to an infinite number of directed line segments, where each of these directed line segments has a different initial point but all of them require the same “change” to change positions from initial point to terminal point. Continuing our example, every change in position between some initial and terminal points in the plane that requires moving two units right and three units down can be represented by the same vector.
We describe a two-dimensional vector in the plane with a pair of numerical components: the change in \(x\) and the change in \(y\text{.}\) If \(P(x_1,y_1)\) and \(Q(x_2,y_2)\) are points in the plane, then the vector associated to the directed line segment \(\abray{PQ}\) has components \(\uvec{v} = (\Delta x,\Delta y) = (x_2-x_1,y_2-y_1)\text{.}\) (A three-dimensional vector in space requires a third component as well: the change in \(z\text{.}\))
Notice what happens when we use the origin \(O(0,0)\) as the initial point and an arbitrary point \(R(x,y)\) as the terminal point in a directed line segment: the vector associated to \(\abray{OR}\) is then \((x-0,y-0) = (x,y)\text{.}\) So when the initial point is the origin, the components of the vector are exactly the coordinates of the terminal point. In Discovery 12.1 we saw that this works in reverse as well. That is, if we have a vector \(\uvec{v} = (v_1,v_2)\text{,}\) then we could consider the numbers \(v_1,v_2\) as coordinates of a point \(R(x,y)\) with \(x=v_1\) and \(y=v_2\text{,}\) and then the vector associated to \(\abray{OR}\) is just \(\uvec{v}\) again. In this way, every vector corresponds to one unique directed line segment with initial point at the origin, and so that is sort of the “natural position” of the vector as a directed line segment. But we will find that it is often convenient to consider other directed line segments that correspond to a particular vector.
We live in a three-dimensional world (or, at least, it appears that way to us), and our little human brains cannot visualize points or arrows in four- or higher-dimensional spaces. However, we can still describe such imaginary objects using our experience from two- and three-dimensional points and vectors. For example, if we had two points \(P\) and \(Q\) in an imaginary four-dimensional space, they would each require four coordinates, so we would describe them as \(P(w_1,x_1,y_1,z_1)\) and \(Q(w_2,x_2,y_2,z_2)\text{.}\) Then the vector corresponding to the directed line segment \(\abray{PQ}\) would have four components and we would compute it as \(\uvec{v} = (\Delta w, \Delta x, \Delta y, \Delta z) = (w_2-w_1,x_2-x_1,y_2-y_1,z_2-z_1)\text{.}\)
Subsection 12.3.2 Vector addition
A vector describes a change in position. If we chain two changes in position together, by making the initial point of the second vector the same as the terminal point of the first vector, then we could consider the overall change in position.
If these are points and vectors in the plane, then clearly the change in position from \(P\) to \(R\) will be described by the total net change in \(x\) and the total net change in \(y\text{,}\) as we discovered in Discovery 12.2. So, in the diagram above, we obtain the components for the vector corresponding to \(\abray{PR}\) by adding corresponding components of \(\uvec{u}\) and \(\uvec{v}\text{.}\) That is, if \(\uvec{u} = (u_1,u_2)\) and \(\uvec{v} = (v_1,v_2)\text{,}\) then the components of the dashed arrow labelled with a question mark are \((u_1+v_1,u_2+v_2)\text{.}\) For obvious reasons, we call this the sum of \(\uvec{u}\) and \(\uvec{v}\text{.}\)
In Discovery 12.2 we also considered the result of interchanging the order of a pair of vectors that have been chained together.
In the diagram above, the vector for \(\abray{PQ'}\) is the same as that for \(\abray{QR}\text{,}\) because they represent the same change in position, just with a different initial point. Accordingly, we have labelled both vectors with \(\uvec{v}\text{.}\) And the same applies to \(\uvec{u}\) with respect to \(\abray{PQ}\) and \(\abray{Q'R}\text{.}\)
The diagram illustrates that if we start at \(P\) and chain together the change-of-position instructions contained in vectors \(\uvec{u}\) and \(\uvec{v}\text{,}\) the order that we do so does not matter — the overall change in position will be from \(P\) to \(R\text{.}\) Thus, the order of vector addition doesn't matter. Algebraically, we could have predicted that this would be the case because it doesn't matter what order you add components: the identities \(u_1 + v_1 = v_1 + u_1\) and \(u_2 + v_2 = v_2 + v_2\) are both valid. But it's useful conceptually to have the above geometric picture of vector addition because, whether you believe this about yourself or not, humans are spatial thinkers. And the geometric version of the vector identity \(\uvec{v} + \uvec{u} = \uvec{u} + \uvec{v}\) makes a pretty picture of a parallelogram, so we call it the parallelogram rule.
For three-dimensional vectors, we can imagine diagrams like the ones we have drawn above floating in space, and the parallelogram rule would hold there as well. In higher dimensions, we cannot draw pictures, but we could imagine that they are similar. At any rate, the algebra of vector addition is the same in any dimension: for \(\uvec{u} = (u_1,u_2,\dotsc,u_n)\) and \(\uvec{v} = (v_1,v_2,\dotsc,v_n)\) in \(\R^n\text{,}\) we have
Subsection 12.3.3 The zero vector
There is one special change in position that is unlike any other — the one where the initial and terminal points are the same, so that there is actually no change in position. In two dimensions, this means there is no change in either \(x\) or \(y\text{,}\) so the components are \((0,0)\text{.}\) Similarly, in any number of dimensions we have the zero vector \(\zerovec = (0,0,\dotsc,0)\text{.}\)
As we explored in Discovery 12.3, if we chain together a vector \(\uvec{v}\text{,}\) representing some change in position, with the zero vector, which represents no change, then the net result is just the change of \(\uvec{v}\text{.}\) That is, \(\uvec{v} + \zerovec = \uvec{v}\text{,}\) and also \(\zerovec + \uvec{v} = \uvec{v}\text{.}\)
Subsection 12.3.4 Vector negatives and vector subtraction
If we move from \(P\) to \(Q\text{,}\) and from there move from \(Q\) back to \(P\text{,}\) the net result is no change in position, which is represented by the zero vector. This means if we add the vector corresponding to \(\abray{PQ}\) to the one corresponding to \(\abray{QP}\text{,}\) the result is \(\zerovec\text{.}\) So if we label the vector for \(\abray{PQ}\) as \(\uvec{v}\text{,}\) it seems reasonable to label the vector for \(\abray{QP}\) as \(-\uvec{v}\text{,}\) the negative of \(\uvec{v}\text{,}\) so that we have \(\uvec{v} + (-\uvec{v}) = \zerovec\text{.}\)
If we are to have \(\uvec{v} + (-\uvec{v}) = \zerovec\text{,}\) and the components of \(\zerovec\) are all \(0\text{,}\) then since we add vectors by adding corresponding components, the components of \(-\uvec{v}\) must be the negatives of the components of \(\uvec{v}\text{.}\) For example, if \(\uvec{v} = (v_1,v_2)\) in the plane, then \(-\uvec{v} = (-v_1,-v_2)\text{.}\) In any dimension, we have
This relationship between the components of \(\uvec{v}\) and \(-\uvec{v}\) will lead to an identity between the negative and a certain scalar multiple of a vector in Subsection 12.3.5 below.
Remembering that the “natural” position for a vector is with its tail at the origin, it's useful to visualize negatives in the following manner.
That is, the negative of a vector will change positions by the same distance but in the opposite direction.
To subtract vectors, we add to a vector the negative of another.
Here, the diagonal vector labelled \(\uvec{u} - \uvec{v}\) is obtained by adding \(\uvec{u}\) and \(-\uvec{v}\text{.}\) As we explored in Discovery 12.5, we get an interesting pattern if we draw in another copy of the vector labelled \(\uvec{u} - \uvec{v}\) with its initial point at \(R\text{.}\)
Triangle \(\abctriangle{RP'P}\) creates the vector addition pattern \(\uvec{u} + (-\uvec{v}) = \uvec{u} - \uvec{v}\text{.}\) But notice that \(\abctriangle{ORP}\) creates a vector addition pattern starting at \(O\) and ending up at \(P\text{,}\) by \(\uvec{v} + (\uvec{u}-\uvec{v}) = \uvec{u}\text{.}\) So we can think of a difference of two vectors as a vector that runs between the heads of the two vectors in the difference when they share the same initial point. Algebraically, we can think of the \(\uvec{v}\) and \(-\uvec{v}\) cancelling in the expression \(\uvec{v} + (\uvec{u} - \uvec{v})\text{,}\) leaving just \(\uvec{u}\text{.}\)
Of course, there are two vectors that run between the heads of \(\uvec{u}\) and \(\uvec{v}\text{,}\) namely \(\uvec{u}-\uvec{v}\) and its negative.
Now \(\abctriangle{PP'R}\) creates a vector addition pattern starting at \(P\) and ending up at \(R\text{,}\) so that \(\uvec{v} + (-\uvec{u}) = \uvec{v} - \uvec{u}\text{.}\) But also \(\abctriangle{OPR}\) creates a vector addition pattern starting at \(O\) and ending up at \(R\text{,}\) so that \(\uvec{u} + (\uvec{v} - \uvec{u}) = \uvec{v}\text{.}\) And finally, the fact that \(\uvec{u} - \uvec{v}\) and \(\uvec{v} - \uvec{u}\) both run between \(P\) and \(R\text{,}\) but in opposite directions, verifies geometrically that \(-(\uvec{u} - \uvec{v}) = \uvec{v} - \uvec{u}\text{,}\) as we would expect algebraically.
Subsection 12.3.5 Scalar multiplication
Geometrically, when we scalar multiply a vector we “stretch” or scale its length by the scale factor. (If this scale factor is negative, then we also flip the vector around in the opposite direction.) Here are some examples.
Notice how each of these vectors either points in the same direction as or in the opposite direction to \(\uvec{v}\text{.}\) In particular, they are all parallel to one another. This happens precisely when the vectors are scalar multiples of one another.
Thinking of the vectors in the diagram above as vectors in the plane, if we scale \(\uvec{v}\) by a factor of \(2\text{,}\) then our knowledge of similar triangles tells us that the change in both \(x\) and \(y\) must be double.
So if \(\uvec{v} = (v_1,v_2)\text{,}\) then \(2\uvec{v} = (2v_1,2v_2)\text{.}\) This relationship between original vector \(\uvec{v}\) and scaled vector \(k\uvec{v}\) holds in general, in any dimension, and even for negative \(k\text{:}\)
In the case that \(k = -1\text{,}\) we obtain the identity \((-1)\uvec{v} = -\uvec{v}\text{,}\) as promised earlier.
Remark 12.3.1.
It may seem redundant to write \((-1)\uvec{v} = -\uvec{v}\text{,}\) don't both sides mean the same thing? In terms of the effect on components of \(\uvec{v}\text{,}\) yes they are the same. However, when we explore abstract vectors in Chapter 16, we won't have components or the geometric notion of “opposite direction” as means of seeing this equality, and so there will initially be a subtle difference between the idea of a vector having an additive negative (so that \(\uvec{v} + (-\uvec{v}) = \zerovec\)) and the operation of scalar multiplying a vector by the particular scalar \(-1\text{.}\)
We can connect scalar multiplication to addition, as we explored in Discovery 12.6. If we add a vector to itself, then the sum vector will be twice as long as the original.
So we have \(\uvec{v} + \uvec{v} = 2 \uvec{v}\text{.}\)
Subsection 12.3.6 Vector algebra
We have already discovered a few rules of vector algebra, such as
In Discovery 12.7, we explored a version of the distributive rule \(k (\uvec{u} + \uvec{v}) = k \uvec{u} + k \uvec{v}\) in the case \(k=2\text{.}\)
We will provide more rules of vector algebra as Proposition 12.5.1 in Subsection 12.5.1. In Discovery 12.12, we decided that the algebra of vectors is the same as the algebra of column matrices (which we have already been referring to as column vectors), so we should be able to anticipate a number of the vector algebra rules that will appear in that proposition.
Warning 12.3.2.
There is no multiplication operation for vectors!
Algebraically, vectors in \(\R^n\) are the same as column matrices, and you cannot multiply two column matrices together because their sizes do not match up (except in \(\R^1\text{,}\) but let's ignore that for now). This also means that you cannot square a vector, you cannot square-root a vector, you cannot invert a vector, and you cannot divide by a vector. Do not try to use any of these operations in vector algebra! In Chapters 13–14, we will encounter some operations tied to the geometry of vectors that we will call “vector products” and for which we will use multiplication-like notation, but they will be for very specific geometric purposes and do not really correspond to our idea of multiplication in the algebra of numbers.
Subsection 12.3.7 The standard basis vectors
In Discovery 12.8, we encountered two very special vectors in the plane, \(\uvec{e}_1 = (1,0)\) and \(\uvec{e}_2 = (0,1)\text{.}\) These two vectors could be considered the fundamental changes in position in the plane — \(\uvec{e}_1\) represents a change by one unit right, and \(\uvec{e}_2\) represents a change by one unit up.
Any change in position can be built out of these two fundamental changes in position. Using the example in Discovery 12.8, the vector \(\uvec{v} = (5,2)\) represents a change in position by \(5\) units right and \(2\) units up. We can achieve the “\(5\) units right” part with \(5 \uvec{e}_1 = (5,0)\) and the “\(2\) units up part” with \(2 \uvec{e}_2 = (0,2)\text{.}\) To get the total change in position represented by \(\uvec{v}\text{,}\) we can combine these two building blocks in the linear combination \(\uvec{v} = 5 \uvec{e}_1 + 2 \uvec{e}_2\text{.}\)
As you can imagine, every vector in the plane can be decomposed into a linear combination of \(\uvec{e}_1\) and \(\uvec{e}_2\) in this manner: for \(\uvec{v} = (v_1,v_2)\text{,}\) we have \(\uvec{v} = v_1\,\uvec{e}_1 + v_2\,\uvec{e}_2\text{.}\) For this reason, the two vectors \(\uvec{e}_1,\uvec{e}_2\) together are called the standard basis vectors in \(\R^2\text{,}\) as they form a basis from which every other vector can be constructed. To use an analogy with chemistry, these two vectors are the basic atoms of \(\R^2\text{,}\) and every other vector in \(\R^2\) is a molecule built out of a specific combination of these atoms. Since there are only two fundamental directions in \(\R^2\) (right/left and up/down), it is not surprising that we need only two basis vectors to represent all possible directions. This is the reason we call vectors in \(\R^2\) two-dimensional vectors.
In \(\R^3\text{,}\) there are three fundamental directions, two horizontal and one vertical. We might use navigational terminology for the two horizontal directions and describe them as north/south and east/west, and then we can refer to the vertical direction as up/down. So we need three standard basis vectors in \(\R^3\text{,}\)
which we can visualize as below.
As before, any vector in \(\R^3\) can be decomposed as a linear combination of these three fundamental vectors. For example, the vector \((1,-1,2)\) decomposes as
And we can repeat all this in \(\R^n\) for any value of \(n\text{,}\) where there are \(n\) standard basis vectors,
In fact, given a vector \(\uvec{v} = (v_1,v_2,\dotsc,v_n)\) in \(\R^n\) (whether \(n=2\) or \(n=3\) or higher), when we try to decompose
we find that there is only one unique combination of scalar values that can fill in the blanks and create an equality between \(\uvec{v}\) on the left and the linear combination on the right: