We have previously considered the matrix-times-vector pattern with a geometric perspective, especially in the context of a transition matrix that achieves a similarity relation, as a way to transform vectors in Rn. We can think of this as an input-output process, where an input vector x is transformed into an output vector Ax. In those past explorations, we focused on the case that A is square, so that input vectors and transformed output vectors are the same dimension, but there is no reason to restrict to just the square case.
As described in the prelude to Discovery 42.2, we once again use the patterns of Rn as a guide to create axioms for abstract vector spaces. A vector space is defined by the operations on its objects, vector addition and scalar multiplication, so we used the interactions of a matrix transformation TA:Rn→Rm with vector addition and scalar multiplication as a model for the abstract idea of a linear transformation between vector spaces. We know matrix multiplication satisfies algebra rules
(Rule 1.c and Rule 2.d of Proposition 4.5.1). Note that in each of these two algebra rules, the type of addition or scalar multiplication on the left-hand side of the equals sign is different from the type of addition or scalar multiplication on the right-hand side. On the left, the operation is being performed between n-dimensional vectors, before multiplication by A. On the right, the operation is being performed between m-dimensional vectors, after multiplication by A.
for all vectors u,v in V and all scalars k, and these two properties are the foundation for the definition of the concept of linear transformation. And just as for the matrix-transformation version of these properties, the type of addition or scalar multiplication on the left-hand side is different from the type of addition or scalar multiplication on the right-hand side. On the left, the operation is being performed between vectors in V,before being input into the transformation T, according to how the operation is defined in V. On the right, the operation is being performed between vectors in W,after being input into the transformation T, according to how the operation is defined in W.
In Discovery 42.2, we also considered the interaction of matrix transformations with other vector concepts, such as linear combinations, negatives, and the zero vector. However, in an abstract vector space, each of these concepts is defined in terms of or relative to the operations of addition and scalar multiplication, so we expect that the two linearity properties (⋆) of an abstract linear transformation will also cause it to satisfy similar properties relative to these other concepts. (See Proposition 42.5.1 in Subsection 42.5.1.)
In Discovery 42.4, we found that knowing the image vectors for each vector in a spanning set for the domain space of a linear transformation T:V→W is enough to recover the image vectors for all input vectors in the domain space.
are known, then the image of every linear combination of the vj can be determined from the corresponding linear combination of the wj. As these vectors form a spanning set for V, the output for each domain vector can be determined in this way. See Example 42.4.7 for an example of using a domain space spanning set to analyze a linear transformation.
Even though a linear transformation is completely determined by its image vectors for a spanning set of the domain space V, it's important to use a basis for V when using these patterns to define a linear combination. The redundancy created by a dependent spanning set (see Discovery 19.2) will also create ambiguity in defining T(vj)=wj, since if one of the vj is a linear combination of the others and the wj are truly arbitrarily chosen, the wj won't necessarily exhibit the same dependence relationship as the vj. See Example 42.4.8 for an example of how this could go wrong if we use a dependent spanning set for the domain space to define a linear transformation.
Suppose we apply Procedure 42.3.1 to the task of creating a linear transformation T:Rn→Rm by choosing the standard basis of the domain space Rn, and then setting
And in light of Discovery 42.4, since the outputs of T and TA agree on the standard basis vectors for the domain space Rn, they will also agree on every linear combination of those standard basis vectors, i.e. they will also agree on every vector in the domain space Rn. That is, T=TA.
In effect, every linear transformation Rn→Rm is a matrix transformation. (See Corollary 42.5.4 in Subsection 42.5.1.) For a given linear transformation T:Rn→Rm, the matrix
so that T=TA for A=[T], is called the standard matrix of T. Expressing the above pattern in words, the standard matrix of a transformation is the matrix whose columns are the image vectors for the standard basis vectors under the transformation.
for all v in V, where 0W is the zero vector in W. If we want to distinguish the zero transformation from the zero vector or the zero matrix or etc., we could write 0V,W:V→W, though this notation is cumbersome and usually context alone is sufficient to determine what 0 means.
Given a vector space V and a scalar a, we can create a scalar multiplication linear transformation ma:V→V by scaling each vector in V by scale factor a:
We considered this operator in Task 42.3.d, where it should have been obvious that ma satisfies the linearity properties (⋆) from vector Axiom S 2 and vector Axiom S 4. Note that the scalar values a=−1,0,1 create other special operators already discussed above:
Suppose B is a basis for an n-dimensional vector space V. We have already seen that the process of determining coordinate vectors relative to B is linear:
The Linearity of inner products implies that by choosing a fixed vector u0 in an inner product space V, we can create a transformation Tu0:V→R1 by pairing with u0:
In Subsection 16.4.2, we created a vector space out of a collection of functions f:D→R defined on some domain D of real numbers by defining vector addition and scalar multiplication via addition and scaling of output values. Linear transformations are first and foremost functions, so we can attempt to do the same with the collection L(V,W) of all linear transformations V→W for vector spaces V,W. As the codomain space W is equipped with vector addition and scalar multiplication operations, we can indeed use the pattern of adding and scaling transformations by adding and scaling outputs:
Axiom A 1 and Axiom S 1 require that a vector space be closed under the addition and scalar multiplication operations, but we leave it to you, the reader, to verify that a sum of linear transformations is also linear, and that a scalar multiple of a linear transformation is also linear. And, in fact, all ten Vector space axioms will be satisfied by the above definitions of addition and scalar multiplication of linear transformations, making L(V,W) into a vector space.
In the case of L(Rn,Rm), we saw in Subsection 42.3.4 that each linear transformation Rn→Rm corresponds to multiplication by some m×n matrix. How does adding and scalar multiplying linear transformations in L(Rn,Rm) affect the corresponding matrices? Suppose T1=TA1 and T2=TA2 for m×n matrices A1,A2. Let B represent the matrix sum A1+A2. Then for each vector x in Rn, we have
Combined, these two properties of the relationship between a linear transformation Rn→Rm and the corresponding m×n matrix mean that, as a vector space, L(Rn,Rm) can be considered as the “same” space as Mm×n(R).
In Discovery 42.6, we considered transformations Rn→R1. Such a transformation is special for two reasons. First, the output is more naturally considered as a number rather than a vector, which is why we refer to such a transformation as a linear functional. Second, the standard matrix of such a linear transformation will be a 1×m matrix, which itself can be considered as a (row) vector in Rn. And multiplication of column vectors in Rn by some fixed row vector is effectively a dot product. That is, every linear functional Rn→R corresponds to dot product against a fixed n-dimensional vector. This creates a correspondence between the dual space (Rn)∗=L(Rn,R1) and the original space Rn, where each transformation in (Rn)∗ is matched with the row vector in Rn which describes the transformation in terms of the dot product.
To try to generalize the pattern of linear functionals on Rn, we expect that if V is a finite dimensional vector space, then there should be a connection between vectors in V and transformations in the dual space V∗=L(V,R1). We saw in Subsection 42.3.5 that such a connection exists if V is an inner product space, but we can do this even if V is not an inner product space by simply choosing a basis for V.
Suppose B={v1,…,vn} is a basis for finite-dimensional, real vector space V. Using the pattern of Proposition 37.5.6 as inspiration, we can create an inner product on V just by pretending B is an orthonormal basis of V, and using the unique expansions
This pairing will be an inner product, and B will indeed by orthonormal with respect to it. Then we can use a fixed vector w0 in V to create a linear functional Tw0 by
But the correspondence can be worked the other way as well, associating to a fixed linear functional f:V→R1 some vector w0 in V so that f=Tw0. Indeed, as in Discovery 42.4, the functional f is completely determined by its outputs on the basis B. Applying f to the basis vectors vj, we obtain scalars
Above we saw that a choosing a basis for a finite-dimensional vector space V creates a correspondence between vectors in V and linear functionals in the dual space V∗=L(V,R1). But V∗ is also a finite-dimensional vector space, and so choosing a basis for V∗ creates a correspondence between vectors in V∗ (i.e. linear functionals on V) and linear functionals in the double dual space V∗∗=L(V∗,R1). We can bridge these two correspondences to create a direct correspondence between vectors in V and vectors in the double dual V∗∗.
Since the individual correspondences V↔V∗ and V∗↔V∗∗ depend on choosing bases for the spaces, they are coordinate-dependent. However, in more advanced study of linear algebra, you might learn that there is actually a coordinate-free correspondence directly between V and its double dual space V∗∗. This correspondence is easy to describe: for v in V, define linear functional T∗v on V∗ by
To make sense of this definition, keep in mind that a linear functional on V∗ should take as inputs linear functionals on V. We will leave further analysis of this correspondence to your future studies in linear algebra.
While we focused on the real case in this section so far, everything works the same for complex vector spaces. Linear transformations Cn→Cm are still matrix transformations, but by complex matrices. We still have zero and identity operators on complex vector spaces, and linear functionals, and so on.