Skip to main content
Logo image

Section 13.3 Concepts

Subsection 13.3.1 Values of uv

In Discovery 13.1, we compared the graph of the cosine function on the domain 0θπ with the formula
(✶)cosθ=uvuv,
where θ is the angle between nonzero vectors u and v. On the right of equation (✶), the denominator is always positive, so whether the whole fraction is positve, negative, or zero depends entirely on the dot product in the numerator. On the left, the cosine function is positive, negative, or zero precisely when the angle θ is acute, obtuse, or right. So we come to the following conclusions.
θ uv
acute: 0θ<π/2 positive
right: θ=π/2 zero
obtuse: π/2<θπ negative
Figure 13.3.1. Values of the dot product for vectors at various angles.

Subsection 13.3.2 Orthogonal vectors

Right angles are extremely important in geometry, and from Figure 13.3.1 we see that the dot product gives us a very convenient way to tell when the angle θ between two nonzero vectors u and v is right: we have θ=π/2 precisely when uv=0. In the plane or in space, u and v will be perpendicular when θ=π/2 and uv=0. Since we can’t “see” right angles and perpendicular lines in higher dimensions, in general we say that u and v are orthogonal when uv=0.

Subsubsection 13.3.2.1 Orthogonal vectors in R2

In Discovery 13.2, we tried to find a pattern to the task of choosing some vector that is orthogonal to a given one in the plane. Rather than struggle with the geometry, we unleash the power of algebra: given vector u=(a,b), we are looking for a vector v so that uv=0. Expanding out the dot product, we are looking to fill in the blanks in the following equation with components for v:
aX+bX=0.
Two numbers add to zero only if one is the negative of the other. We can make both terms in the sum the same number by entering b in the first blank and a in the second, so we can make the sum cancel to zero by also flipping the sign of one of those entries. For example,
ab+b(a)=0.
We have now answered the question in Discovery 13.2.c.
Pattern 13.3.2. Orthogonal vectors in the plane.
Given vector u=(a,b) in the plane, two examples of vectors that are orthogonal to u are v=(b,a) and v=(b,a), and every vector that is orthogonal to u is some scalar multiple of this example v.

Subsection 13.3.3 Orthogonal projection

Orthogonal projection is a vector solution to a problem in geometry.

Question 13.3.4.

Given a line through the origin in the plane, and a point not on the line, what point on the line is closest to the given point?
In Question 13.3.4, write for the line through the origin and Q for the point not on that line. Consider the point P on at the foot of the perpendicular to from Q . Any other point P on will form a right triangle with P and Q, making it farther from Q than P, since the distance PQ is the length of the hypotenuse in the right triangle.
described in detail following the image
Diagram illustrating the shortest distance from a point to a line in the plane.
All we know about P is that it is on line and it is at the vertex of a right angle with and Q. But if we introduce some vectors to help tackle this problem, then maybe we can use what we know about the dot product and right angles to help determine P.
described in detail following the image
Diagram illustrating orthogonal projection onto a line through the origin.
In this diagram, u is the vector corresponding to directed line segment OQ, and p is the vector corresponding to the directed line segment OP, where P is our unknown closest point. Since p is placed with its tail at the origin, the components of p are precisely the coordinates of P. So determining p will solve the problem.
We are assuming that the line is known, and it would be nice to also have a vector means of describing it. But the vectors created by the points on this line (using the origin as a universal tail point) will all be parallel to each other, so (as we discovered in Discovery 13.3.a) line could be described as all scalar multiples of a particular vector a. This vector can be arbitrarily chosen as any vector parallel to the line. Once we have chosen a, we have reduced our problem from determining the two unknown components of the vector p to determining a single unknown scalar k so that p=ka.
As mentioned, since P is the closest point, the directed line segment PQ must be perpendicular to . On the diagram above, we have used the vector n to represent this direct line segment. As in Discovery 13.3.b, we know that na must be zero — this is the perpendicular condition. However, the vector n is unknown as well, since we don’t know its initial point. But we can also use the triangle formed by u, n, and p to replace n:
p+n=un=up=uka
Replacing n by this expression in the condition na=0 gives us an equation of numbers that we can solve for the unknown scale factor k, as we did in Discovery 13.3.d:
k=uaa2.
This vector p=ka pointing from the origin to the desired closest point P is called the projection of u onto a or sometimes the vector component of u parallel to a, and we write projau to represent it.

Remark 13.3.6.

It is not actually necessary that Q be external to the line. If you were to carry out the procedure above in the case that Q lies on , the calculations would end up with p=u, confirming that Q was already the point on the line that is closest to itself.
The normal vector n in the diagram above is sometimes called the vector component of u orthogonal to a. Together, the projection vector and corresponding normal vector are called components of u (relative to a) because they represent an orthogonal decomposition of u:
u=p+n,
where p is parallel to a and n is orthogonal to a. While this decomposition is relative to a, it is really only the direction of a that matters — if a is parallel to a (even possibly opposite to a), then both
p=projau=projau,n=up=uprojau=uprojau
will be true.

Remark 13.3.8.

  1. These procedures and calculations can be easily modified to work for lines that do not pass through the origin: simply choose some arbitrary “initial” point R on the line to “act” as the origin.
  2. All of these calculations can be performed in higher dimensions as well. In higher dimensions, it is true that there is no longer one unique perpendicular direction to a given vector a, but the calculation of n as above will pick out the correction direction to extend from the line to the point Q at a right angle to the line.

Subsection 13.3.4 Normal vectors of lines in the plane

Consider the line 2x+3y=0 that we investigated in Discovery 13.6. The point (3,2) is on this line, since
(✶✶)23+3(2)=0.
The left-hand side of this calculation looks a lot like a dot product — we could reinterpret equation (✶✶) as
(2,3)(3,2)=0.
So verifying that the point (3,2) is on the line is equivalent to checking that the corresponding vector v=(3,2) (with its tail at the origin) is orthogonal to the vector n=(2,3) whose components are the coefficients from our line equation.
described in detail following the image
Diagram of a normal vector to a line in the plane (homogeneous case).
Every other point x=(x,y) on the line satisfies the same relationship, as the equation for the line could be rewritten in a vector form as
(✶✶✶)nx=0.
The vector n is called a normal vector for the line. Note that normal vectors for a line are not unique — every nonzero scalar multiple of n will also be normal to the line, and this is equivalent to noting that we could multiply the equation 2x+3y=0 by any nonzero factor to obtain a different equation that represents the same line in the plane.
In Discovery 13.7 we considered a line defined by a nonhomogeneous equation 2x+3y=8. This line has the same slope as the line defined by 2x+3y=0 that we investigate above, and so the vector n=(2,3) obtained from the coefficients on x and y in the equation must still be normal. The constant 8 just changes the y-intercept.
described in detail following the image
Diagram of a normal vector to a line in the plane (nonhomogeneous case).
In the homogeneous case, vectors from the origin determined by a point on the line were also parallel to the line. Since things have shifted away from the origin in the nonhomogeneous case, to get a vector parallel to the line we need to consider two vectors from the origin to points on the line. Two convenient points for this the line are Q(1,2) and R(4,0), with corresponding vectors x0=(1,2) and x1=(4,0). Then the difference vector
v=x1x0=(3,2)
is parallel to the line, as in the diagram above. In fact, this vector v is the same as previous vector v that appears parallel to the line through the origin in the diagram for the homogeneous case above, so we know it satisfies nv=0.
Is there a way to use the normal vector n to create a vector condition by which we can tell if a vector x represents a point on the line, as we did with equation (✶✶✶) in the homoegenous case? We need two points on the line to create a parallel difference vector, but we could compare the variable vector x with a arbitrarily chosen fixed vector representing a point on the line (like x0, say).
described in detail following the image
Diagram of a normal vector to a line in the plane (nonhomogeneous case).
Every such difference vector xx0 is parallel to the line and hence orthogonal to the normal vector n, so that we can describe the line as all points where the corresponding vector x satisfies
(†)n(xx0)=0.
This is called the point-normal form for the line, referring to the point on the line at the terminal point of x0 and the normal vector n.

Pattern 13.3.9. Point-normal form for a line in R2.

If (x0,y0) is a point on the line :ax+by=d (that is, ax0+by0=d is true), then can alternatively be described as all points (x,y) that satisfy
(a,b)((x,y)(x0,y0))=0.

Remark 13.3.10.

It may seem like the line parameter d has disappeared in converting from algebraic form ax+by=d to point-normal form. But it has merely be replaced by the point (x0,y0), since d=ax0+by0. In fact, if we use the algebraic properties of the dot product to expand the left-hand side of the point-normal form equation, we can recover the original algebraic equation:
(a,b)((x,y)(x0,y0))=0(a,b)(x,y)(a,b)(x0,y0)=0(ax+by)(ax0+by0)=0(ax+by)d=0ax+by=d.

Subsection 13.3.5 Normal vectors of planes in space

A similar analysis can be made for an equation ax+by+cz=d describing a plane in space. The coefficients form a normal vector n=(a,b,c). For vectors x0 and x1 that both have initial point at the origin and terminal points on the plane, then the difference vector x1x0 is parallel to the plane, hence normal to n. If we keep a fixed choice of x0 but replace x1 by a variable vector x, we can describe the plane as all points whose difference is orthogonal to n, giving us a point-normal for a plane just as in equation (†).

Pattern 13.3.11. Point-normal form for a plane in R3.

If (x0,y0,z0) is a point on the plane Π:ax+by+cz=d (that is, ax0+by0+cz0=d is true), then Π can alternatively be described as all points (x,y,z) that satisfy
(a,b,c)((x,y,z)(x0,y0,z0))=0.

Remark 13.3.12.

A line in space does not have a point-normal form, because it does not have one unique normal “direction” like a line in the plane or a plane in space does. To describe a line in space in a similar fashion you would need two normal vectors. We will see several more convenient ways to describe a line in space in the next chapter.

Subsection 13.3.6 The cross product

Seeing how the algebraic equation for a plane in R3 is connected to a normal vector to the plane, a basic problem is how to quickly obtain a normal vector. If we know two vectors that are parallel to the plane in question, the problem reduces to the following.

Question 13.3.13.

Given two nonzero, nonparallel vectors in R3, determine a third vector that is orthogonal to each of the first two.
described in detail following the image
Diagram of the setup for the cross product problem (Question 13.3.13 above).
So if u=(u1,u2,u3) and v=(v1,v2,v3) are our starting vectors, we would like to simultaneously solve the equations
ux=0,vx=0,
for the unknown vector x=(x,y,z). Expanding out the dot products, we get (surprise!) a system of linear equations:
{u1x+u2y+u3z=0,v1x+v2y+v3z=0.
Specifically, we get a homogeneous system of two equations in the three unknown coordinates x,y,z. Now, since this system is homogeneous, it is consistent. But its general solution will also require at least one parameter, since its rank is at most 2, while we have three variables. In the diagram above, we can see what the “freedom” of a parameter corresponds to — we can make x longer or shorter, or turn it around to be opposite of the way it is pictured, and it will remain orthogonal to u and v. Our end goal is a calculation formula and procedure that will compute one particular solution to this problem, so let’s introduce a somewhat arbitrary additional equation to eliminate the need for a parameter in the solution.
{x+y+z=1,u1x+u2y+u3z=0,v1x+v2y+v3z=0.
In matrix form, this system can be expressed as Ax=b, with
(††)A=[111u1u2u3v1v2v3],b=[100].
Assuming that detA0, Cramer’s rule tells us the solution to this system.
x=1detA|1110u2u30v2v3|y=1detA|111u10u3v10v3|z=1detA|111u1u20v1v20|=1detA|u2u3v2v3|=1detA|u1u3v1v3|=1detA|u1u2v1v2|
Now, each of x,y,z has a common factor of 1/detA, and all this common factor does is scale the length of our solution vector x without affecting orthogonality with u and v. Even worse, detA depends on that extra equation we threw in, and we would like our solution to depend only on u and v. So let’s remove it and use solution
x=(|u2u3v2v3|,|u1u3v1v3|,|u1u2v1v2|).
We call this the cross product of u and v, and write u×v instead of x. There is a trick to remembering how to compute the cross product: if we replace the top row of A by the standard basis vectors i,j,k in R3, then the cross product will be equal to its determinant expanded by cofactors along the first row. That is, setting
(†††)u×v=|ijku1u2u3v1v2v3|
and expanding the determinant along the first row yields
u×v=|u2u3v2v3|i|u1u3v1v3|j+|u1u2v1v2|k,
as desired. See Example 13.4.4 in Subsection 13.4.3 for an example of using formula (†††) to compute cross products.
The cross product follows the right-hand rule — if you orient your right hand so that your fingers point in the direction of u and curl towards v, then your thumb will point in the direction of u×v.
Computing v×u instead of u×v should still produce a vector that is orthogonal to both u and v, but the right-hand rule tells that the two should be opposite to each other. From equation (†††) we can be even more specific. Computing v×u would swap the second and third rows of the special matrix in equation (†††), and we know that the resulting determinant would be the negative of that for the matrix for computing u×v, and so
v×u=u×v.
See Proposition 13.5.5 in Subsection 13.5.3 for more properties of the cross product.

Remark 13.3.14.

There is one more thing to say about our development of the cross product — Cramer’s rule can only be applied if detA is not zero, where A is the matrix in (††). However, the coefficients in the extra equation we introduced did not figure into our final solution. So if detA ended up being zero for some particular vectors u and v, we could just change the variable coefficients in that extra equation (but keep the 1 in the equals column) so that detA is not zero, and we would still come to the same formula for u×v. And it follows from concepts we will learn in Chapter 20 that it is always possible to fill in the top row of this matrix A so that its determinant is nonzero, as long as we start with nonparallel vectors u and v.