Section 37.3 Concepts
In this section.
Right angles are particularly important in geometry. In the geometry of \(\R^n\text{,}\) a right angle indicates two vectors that “point” in completely independent directions. In Chapter 14, we found that a right angle between vectors corresponded to a dot product of zero. Algebraically, this zero value has many conveniences, so we will borrow the concept of orthogonal vectors for use in every inner product space — even in complex ones, where we have declined to define what a general angle between vectors should be.
Subsection 37.3.1 Orthogonal complements
We have already encountered the concept of complement subspace in Subsection 28.6.3 (see Proposition 28.6.6). In our first explorations of Discovery guide 37.1, we considered a very specific kind of complement \(\orthogcmp{U}\text{,}\) made up of all vectors orthogonal to a given subspace \(U\text{,}\) called the orthogonal complement of \(U\text{.}\)
And we have already considered this concept geometrically in the cases of \(\R^2\) and \(\R^3\) in Chapter 14. As we reminded ourselves in Discovery 37.1, the orthogonal complement of a plane (through the origin) in space is just the normal line to the plane, and symmetrically the orthogonal complement of a line (through the origin) in space is just the plane to which the line is normal. This symmetric relationship between planes and their normal lines suggests a general pattern of \(\orthogcmp{(\orthogcmp{U})} = U\text{,}\) which we will confirm in Corollary 37.5.20.
An orthogonal complement is always a subspace of the inner product space. Note that this fact does not depend on the initial collection \(U\) of vectors actually being a subspace — for every collection \(X\) of vectors in an inner product space, the collection \(\orthogcmp{X}\) of all vectors that are orthogonal to every vector in \(X\) will form a subspace as well. (See Proposition 37.5.13.) But in the case that \(U\) is a subspace, then the complement part of orthogonal complement is intentional: the pair of subspaces \(U, \orthogcmp{U}\) will always form a complete set of independent subspaces in a finite-dimensional inner product space. (See Corollary 37.5.19.)
As we explored in Discovery 37.2, the Linearity of inner products implies that when \(U\) is a subspace, it is enough to check inclusion in \(\orthogcmp{U}\) by checking orthogonality against each vector in some basis for \(U\text{.}\) We saw in Discovery 37.3 how this fact can be used to set up a homogeneous system that will lead from a basis for \(U\) to a basis for \(\orthogcmp{U}\text{.}\) (Also see the examples in Subsection 37.4.1.)
Subsection 37.3.2 Expansion relative to an orthogonal basis
An orthogonal basis for an inner product space is one where each vector in the basis is orthogonal to every other vector in the basis. We are already well-familiar with the concept of orthogonal basis from our experience in \(\R^n\) — the standard basis is always an orthogonal basis (actually, an orthonormal basis) when using the standard inner product.
A basis affords a unique expression (or “expansion”) for each vector in a vector space as a linear combination of basis vectors (Theorem 19.5.3). In Discovery 37.5, we found that the zero result of an inner product for orthogonal vectors, along with the the Linearity of inner products, leads to a pattern for the coefficients in the expansion for a vector relative to an orthogonal basis
in a finite-dimensional inner product space:
Warning 37.3.1.
Because of the conjugate-symmetry of complex inner products (Axiom CIP 1), the order \(\inprod{\uvec{v}}{\uvec{e}_j}\) in each scalar numerator in the general expansion above is important in order to maintain consistency between both the real and complex contexts.
If \(\basisfont{B}\) is an orthonormal basis for an inner product space, then each scalar denominator \(\norm{\uvec{e}_j}\) in the expansion above is equal to \(1\text{,}\) so an expansion becomes even simpler:
Subsection 37.3.3 The Gram-Schmidt orthogonalization process
The point of a basis for a vector space is to have a means to uniquely describe all infinity of vectors in the space as linear combinations of a finite number of basis vectors. We just saw in Subsection 37.3.2 that, relative to an orthogonal basis for a finite-dimensional inner product space, determining that linear combination for a given vector is particularly easy — no row reducing necessary, just compute some inner products and norms. So being able to produce an orthogonal basis for an inner product space is of particular value.
In Discovery 37.6, we attempted to reinvent a procedure by which we can convert a basis into an orthogonal one. (And then, if desired, into an orthonormal one by normalizing each orthogonal basis vector.)
Procedure 37.3.2. Gram-Schmidt orthogonalization.
Given a basis
of a finite-dimensional inner product space, to construct an orthogonal basis
for that space, calculate
If an orthonormal basis is desired, each of the \(\uvec{e}_j\) can be normalized to unit vectors.
Remark 37.3.3.
If aiming to produce an orthogonormal basis, the \(\uvec{e}_j\) can be normalized at any point in the process — either all at the end, or one at a time as they are produced. Both choices have pros and cons.
We will verify that the procedure creates an orthogonal basis in Subsection 37.5.2. For now, consider how each step in the procedure follows from the previous. We will use diagrams as if these were vectors in \(\R^n\text{,}\) as we did in Discovery 37.6.b, to see the geometric pattern in the steps.
We set \(\uvec{e}_1\) to be \(\uvec{v}_1\) because there aren't yet any other vectors with which \(\uvec{e}_1\) needs to be orthogonal (as we haven't yet specified any of the other \(\uvec{e}_j\) vectors). Next, we know from the discussion in Subsection 37.3.2 that if we are able to succeed in creating the orthogonal basis \(\basisfont{B}\text{,}\) then the coefficient on \(\uvec{e}_1\) in the expansion of \(\uvec{v}_2\) relative to \(\basisfont{B}\) is
In other words, the portion of \(\uvec{v}_2\) parallel to \(\uvec{e}_1\) is
This diagram should look familiar — it is essentially the same diagram that we used to illustrate orthogonal projection in \(\R^2\text{.}\) The formula for orthogonal projection even appears to be the same:
And from our experience with orthogonal projection onto a line, the vector running from the head of \(\widetilde{\uvec{v}}_2 \) to the head of \(\uvec{v}_2\) should be both a difference vector
and a normal vector to the line spanned by \(\uvec{e}_1\text{.}\)
We can also visualize this normal vector stemming from the origin.
Since we wanted \(\uvec{e}_2\) to be orthogonal to \(\uvec{e}_1\text{,}\) it would appear from the diagram that we have succeeded in that goal.
The next step follows the same pattern, but now the pair \(\uvec{e}_1,\uvec{e}_2\) span a plane. Similarly to the last step, the vector
represents the portion of \(\uvec{v}_3\) that lies parallel to that plane.
The vector \(\widetilde{\uvec{v}}_3\) also appears to be some sort of “orthogonal projection”, but this time onto a plane instead of a line. (While we have only studied orthogonal projection onto a line in \(\R^n\) up to this point, we will study orthogonal projection more generally in the next chapter.) And the difference vector \(\uvec{e}_3 = \uvec{v}_3 - \widetilde{\uvec{v}}_3\) appears to normal to the plane, hence orthogonal to \(\uvec{e}_1\) and \(\uvec{e}_2\text{,}\) as desired.
Once again, we can also visualize this normal vector as stemming from the origin.
We of course cannot draw a picture of the next step in the Gram-Schmidt process, since that picture would need to be in four dimensions. But we can imagine that the pattern continues, with each new \(\uvec{e}_j\) equal to the difference between the corresponding \(\uvec{v}_j\) and the “orthogonal projection” of \(\uvec{v}_j\) onto the subspace spanned by \(\uvec{e}_1,\dotsc,\uvec{e}_{j-1}\text{.}\) That difference vector will be “normal” to that subspace, hence orthogonal to each of \(\uvec{e}_1,\dotsc,\uvec{e}_{j-1}\text{.}\)
There is one last issue to address. The Gram-Schmidt process is supposed to create a basis for the inner product space, but all we are confident of at this point is that it creates an orthogonal set of \(n\) vectors, where \(n\) is the dimension of the space. However, in Discovery 37.4 we verified that orthogonal vectors are always independent, and having the “right number” of independent vectors is enough to conclude that we have a basis (Corollary 20.5.6).
Subsection 37.3.4 Orthogonal complements from an orthogonal basis
In Discovery 37.8.a, we explored applying the Gram-Schmidt orthogonalization process using the basis of a subspace as the starting point, instead of a basis for the whole inner product space.
Suppose \(U\) is a subspace of an inner product space, with basis \(\basisfont{B}_U = \{ \uvec{u}_1, \dotsc, \uvec{u}_m \} \text{.}\) When we apply the orthogonalization process to just this basis, we start by setting
so \(\uvec{e}_1\) is in \(U\text{.}\) Then we set
a linear combination of vectors in \(U\text{.}\) So \(\uvec{e}_2\) is also in \(U\text{.}\) And so on: at each step the next vector \(\uvec{e}_j\) is a linear combination of a vector from \(\basisfont{B}_U\) and the previous vectors \(\uvec{e}_1,\dotsc,\uvec{e}_{j-1}\text{.}\) If each of those previous vectors is in \(U\text{,}\) then so will the next be.
So the end result will be an orthogonal set of vectors in \(U\) which contains the same number of vectors as \(\basisfont{B}_U\text{.}\) Since orthogonal vectors are independent, this is enough to conclude that the Gram-Schmidt process, applied to a basis for a subspace, will produce an orthogonal basis for that subspace.
As in Discovery 37.8.b, let's take this one step further. The basis \(\basisfont{B}_U\) for \(U\) can be enlarged to a basis
for \(V\) (Statement 1 of Proposition 20.5.8). As we apply the Gram-Schmidt process to \(\basisfont{B}_V\text{,}\) the first \(m\) steps will not involve the \(\uvec{v}_j\text{,}\) and we will effectively be applying the process to \(\basisfont{B}_U\) to obtain an orthogonal basis
of \(U\text{.}\) We then continue the process to obtain an orthogonal basis
for \(V\text{.}\) But then each of these new \(\uvec{f}_j\) vectors is orthogonal to each of the previous \(\uvec{e}_i\text{.}\) Using our thinking from Discovery 37.2, this means that each \(\uvec{f}_j\) is in \(\orthogcmp{U}\text{.}\) And in fact, the collection of \(\uvec{f}_j\) will form an orthogonal basis for \(\orthogcmp{U}\).
We can say that every orthogonal complement pair \(U,\orthogcmp{U}\) in a finite-dimensional inner product space occurs this way. That is, we can create an orthogonal complement pair \(U,\orthogcmp{U}\) by taking an orthogonal basis
for the whole space \(V\) and choosing an index \(\ell\) at which to split it in two:
As indicated by the subscripts, the second “half” of the original set of basis vectors will form an orthogonal basis for the orthogonal complement of the subspace that the first “half” of original basis vectors span: