Skip to main content

Section 2.5 Vector Spaces II

In our first section on vector spaces, we introduced the very important notion of a basis. However, we would also like to discuss subspaces of a vector space and give coordinates on them. This is the concept of linear independence which will be introduced in the first subsection. This concept then is used to define the dimension of a vector space. We then discuss the annihilator of a vector subspace which appears in the dual space.

Subsection 2.5.1 Independence and Dimension

For that we need the following definition.

Definition 2.5.1.

Let \(V\) be a vector space over \(K\text{.}\) We say that the vectors \(\mb{v}_1, \ldots, \mb{v}_n\) are linearly independent if the equality
\begin{equation*} a_1 \mb{v}_1 + \cdots + a_n \mb{v}_n = \mb{0} \end{equation*}
implies that \(a_1, \ldots , a_n\) are all zero.
The equation in Definition 2.5.1 is called a linear relation. So an alternative way of saying that the vectors are linearly independent is that there is no non-trivial linear relation. Yet another way of expressing linear independence is given in the following proposition which the reader can check.
One practical reason to want a linearly independent set of vectors is to parameterize (or ‘give coordinates’ to) their span. Indeed, if \(\mathcal{A} = \{\mb{v}_1 , \ldots, \mb{v}_n \}\) is an ordered set of vectors in \(V\text{,}\) then define the function
\begin{equation*} \chart{}{\mathcal{A}} : K^n \to V \end{equation*}
by
\begin{equation} \left[ \begin{matrix} a_1 \\ \vdots \\ a_n \end{matrix} \right]_{\mathcal{A}} = a_1 \mb{v}_1 + \cdots + a_n \mb{v}_n. \tag{2.5.1} \end{equation}
Then a third way of expressing linear independence can be given as follows.
Note that if \(\mathcal{A}\) is also a basis of \(V\) (i.e. it spans all of \(V\)), then \(\chart{}{\mathcal{A}}\) is the inverse function to \(\coord{}{\mathcal{A}}\text{.}\)
There are instances when it is nice to be able to parameterize lines, planes and higher dimensional subspaces in a given vector space, and \(\chart{}{\mathcal{A}}\) can often be used for this purpose. But to do this in \(V = K^m\text{,}\) one must first see that a given set of column vectors is linearly independent.

Proof.

To see this, just recall from Example 2.2.12 the fact that any linear combination of the \(\mb{c}_i\)’s can be written as the product
\begin{equation*} \left[ \begin{matrix} | \amp | \amp \amp | \\ \mb{c}_1 \amp \mb{c}_2 \amp \cdots \amp \mb{c}_n \\ | \amp | \amp \amp | \end{matrix} \right] \left[ \begin{matrix} a_1 \\ a_2 \\ \vdots \\ a_n \end{matrix} \right] = a_1 \mb{c}_1 + a_2 \mb{c}_2 + \cdots a_n \mb{c}_n. \end{equation*}
In particular, the numbers \(a_1, \ldots, a_n\) give a linear relation
\begin{equation*} a_1 \mb{c}_1 + a_2 \mb{c}_2 + \cdots a_n \mb{c}_n = \mb{0} \end{equation*}
if and only if they solve the matrix equation
\begin{equation*} A \mb{x} = \mb{0}. \end{equation*}
But Theorem 2.4.7 Case (2) says that the solution is unique if and only if there are no free variables of the reduced row echelon equation \(A^\prime \mb{x} = \mb{0}\text{.}\) This implies there are no free columns of \(A\text{.}\)
Proposition 2.5.4, along with row reduction, gives a computational way to check linear independence.

Example 2.5.5. Testing linear independence by solving a matrix equation.

Let us see whether the vectors
\begin{equation*} \threevec{1}{3}{-1} , \threevec{0}{2}{1}, \threevec{2}{4}{-3} \end{equation*}
are linearly independent or not. We compute the reduced row echelon form to see if every column is basic:
\begin{align*} \left[ \begin{matrix} 1 \amp 0 \amp 2 \\ 3 \amp 2 \amp 4 \\ -1 \amp 1 \amp -3 \end{matrix} \right] \amp \stackrel{\mb{r}_2 - 3 \mb{r}_1}{\longrightarrow} \left[ \begin{matrix} 1 \amp 0 \amp 2 \\ 0 \amp 2 \amp -2 \\ -1 \amp 1 \amp -3 \end{matrix} \right], \\ \amp \stackrel{\mb{r}_3 + \mb{r}_1}{\longrightarrow} \left[ \begin{matrix} 1 \amp 0 \amp 2 \\ 0 \amp 2 \amp -2 \\ 0 \amp 1 \amp -1 \end{matrix} \right],\\ \amp \stackrel{1/2 \mb{r}_2}{\longrightarrow} \left[ \begin{matrix} 1 \amp 0 \amp 2 \\ 0 \amp 1 \amp -1 \\ 0 \amp 1 \amp -1 \end{matrix} \right],\\ \amp \stackrel{\mb{r}_3 - \mb{r}_2}{\longrightarrow} \left[ \begin{matrix} 1 \amp 0 \amp 2 \\ 0 \amp 1 \amp -1 \\ 0 \amp 0 \amp 0 \end{matrix} \right]. \end{align*}
We can see that the third column gives a free variable so our set is not linearly independent (also called linearly dependent).
However, by simply forgetting about the third vector (or column in the computation) we see that the first two vectors
\begin{equation*} \threevec{1}{3}{-1} , \threevec{0}{2}{1} \end{equation*}
are linearly independent. Being linearly independent depends heavily on the collection of vectors you are considering.

Proof.

By Proposition 2.5.4 there have to be no free columns of the \(m \times n\) matrix
\begin{equation*} A = \left[ \begin{matrix} | \amp | \amp \amp | \\ \mb{c}_1 \amp \mb{c}_2 \amp \cdots \amp \mb{c}_n \\ | \amp | \amp \amp | \end{matrix} \right]. \end{equation*}
Suppose \(A^\prime\) is the matrix \(A\) in reduced row echelon form. Then \(A^\prime\) has a leading coefficient in every column (otherwise that column corresponds to a free variable). But this gives at least \(n\) leading coefficients for \(n\) unique rows. So the number of rows \(m\) must be at least \(n\) or, in other words, \(n \leq m\text{.}\)
Another important application of Proposition 2.5.4 leads directly to the definition of dimension.

Proof.

Suppose \(V\) has bases \(\mathcal{B} = \{\mb{v}_1, \ldots , \mb{v}_n \}\) and \(\mathcal{C} = \{ \mb{w}_1, \ldots, \mb{w}_m \}\text{.}\) Then by exercise Exercise 2.3.4, we have that there are isomorphisms
\begin{align*} \coord{}{\mathcal{B}} : V \amp \to K^n \amp \coord{}{\mathcal{C}} : V \amp \to K^m. \end{align*}
Now, since \(\mathcal{B}\) is a basis, it is a linearly independent set. And since \(\coord{}{\mathcal{C}}\) is one-to-one, the vectors \(\coord{\mb{v}_1}{\mathcal{C}}, \ldots , \coord{\mb{v}_n}{\mathcal{C}}\) must also be linearly independent (why?). But by the previous corollary, this implies that \(n \leq m\text{.}\) Using the same argument of the vectors \(\coord{\mb{w}_1}{\mathcal{B}}, \ldots , \coord{\mb{w}_m}{\mathcal{B}}\) in \(K^n\) shows that \(m \leq n\text{.}\) Together this gives \(m = n\text{.}\)
This tells us that, while there are infinitely many bases of a vector space, all of them have the same number of vectors in them. This allows us to make the following definition.

Definition 2.5.8.

Given a vector space \(V\) over \(K\text{,}\) we say that the dimension of \(V\) is \(n\) if there is a basis of \(V\) with \(n\) elements. This is denoted \(\dim_K (V) = n\text{.}\) If \(V\) has no finite basis, we say that \(V\) is infinite dimensional.
We record a nice fact that uses our knowledge of linear independence.

Proof.

Suppose \(V\) is \(n\) dimensional. Then by Corollary 2.5.6 and the argument above, any linearly independent set in \(V\) has at most \(n\) vectors. Let \(\mathcal{A} = \{\mb{w}_1, \ldots, \mb{w}_k\}\) be a set of linearly independent vectors in \(W\) of maximal size. Then we contend that \(\mathcal{A}\) is a basis of \(W\text{.}\) If not, then there is a \(\mb{w} \in W\) which is not in the span of \(\mathcal{A}\text{.}\) But then we see that \(\{\mb{w} , \mb{w}_1, \ldots, \mb{w}_k\}\) is also linearly independent (why?). This contradiction shows that \(W\) must have a basis.
Now, as \(W\) is a proper subspace of \(V\text{,}\) there is a vector \(\mb{v}\) of \(V\) which is not in \(W\text{.}\) But then, for the same reason as above, the set \(\{\mb{v}, \mb{w}_1, \ldots, \mb{w}_k\}\) must be linearly independent. By Corollary 2.6.4, this means \(k + 1 \leq n\) which implies \(k \lt n\text{.}\)
Instead of considering linear independence of rows and columns, we can focus on the span of the rows and/or columns.

Definition 2.5.10.

Given an \(m \times n\) matrix \(A\) with entries in \(K\text{,}\)
  1. the column space of \(A\) is the span of the columns of \(A\) in \(K^m\text{.}\)
  2. the row space of \(A\) is the span of the rows of \(A\) in \(K^n\text{.}\)
Note that Example Example 2.2.12 shows that the column space is equal to the image of the linear transformation obtained by left multiplying a column vector in \(K^n\) by \(A\) while the row space is the image of right multiplying a row vector in \(K^m\text{.}\) The following proposition gives bases for both of these spaces.

Proof.

To see that \(\mathcal{C}\) is a basis, note that it is linearly independent by Proposition 2.5.4. Indeed, the reduced row echelon form of the matrix with only these columns is the same as the submatrix of \(A^\prime\) with only these columns. In particular, it has no free columns and the cited proposition implies that \(\mathcal{C}\) is linearly independent. On the other hand, adding any other column of \(A\) to \(\mathcal{C}\) fails the test in Proposition 2.5.4 for the same reason. But that means that any other column can be written as a linear combination of the columns in \(\mathcal{C}\) (why?). Thus \(\mathcal{C}\) spans the column space and is linearly independent. By Proposition 2.5.2, \(\mathcal{C}\) is a basis for the column space.
For the row space, we note that all of the row vectors of \(\mathcal{R}\) are linear combinations of the rows in \(A\) because row operations create rows that are linear combinations of the existing rows. Also, since row operations are reversible (in other words, \(A\) can be recovered by performing row operations on \(A^\prime\)), every row of \(A\) is a linear combination of those in \(A^\prime\text{.}\) This implies the span of \(\mathcal{R}\) is the row space (why?). On the other hand, if
\begin{equation*} a_1 \mb{r}^\prime_1 + \cdots + a_s \mb{r}^\prime_s \end{equation*}
is a linear combination, then projecting to the \(i_1, i_2, \ldots, i_s\) basic columns, we have the vector
\begin{equation*} \left[ \begin{matrix} a_1 \amp a_2 \amp \cdots \amp a_s \end{matrix} \right] \end{equation*}
This is because the coordinate of \(\mb{r}_j^\prime\) in the \(i_j\) column is \(1\) and is zero for all other \(i_k\) columns (because \(A^\prime\) is in reduced row echelon form). Thus the linear combination is the zero vector only if all \(a_i\) are zero, meaning \(\mathcal{R}\) is linearly independent. Again by Proposition 2.5.2 we have that \(\mathcal{R}\) is a basis.
We note here that the bases for the column and row spaces have the same number of vectors! This motivates the following definition.

Definition 2.5.12.

The rank of \(A\) is the dimension of the row space of \(A\text{,}\) or equivalently, the dimension of the column space of \(A\text{.}\)
The student should now be able to understand some of jargin in the Sage description of vector subspaces. Indeed, the methods .row_space() and .column_space() can be used to define the row and column space of a matrix.
The ‘degree’ in this description is the dimension of the ambient space (i.e. the number of columns), and the rank is the dimension of the row space. Indeed, using the method .dimension() the following code checks that the rank of \(A\) also equals the dimension of the column space.

Subsection 2.5.2 Annihilators

In the previous section on vector spaces, we defined the dual space \(V^*\) of a vector space \(V\text{.}\) This construction can be extended to vector subspaces. Furthermore, there is a new notion that can be given.

Definition 2.5.13.

Given a vector subspace \(U \leq V\) the annihilator of \(U\) is the subspace,
\begin{equation*} U^\circ = \left\{ \delta \in V^* : \delta (\mb{u} ) = \mb{0} \text{ for all }\mb{u} \in U \right\}. \end{equation*}
This construction satisfies some basic properties.

Proof.

For the first two statements, let \(\mathcal{B} = \{\mb{u}_1, \ldots, \mb{u}_m, \mb{v}_{m + 1}, \ldots, \mb{v}_n\}\) be a basis for \(V\) such that \(\{\mb{u}_1, \ldots, \mb{u}_m\}\) is a basis of \(U_1\text{.}\) Then it is an exercise to see that \(\{\mb{v}_1^*, \ldots, \mb{v}_{n - m}^* \}\) is a basis of \(U_1^\circ\text{.}\) This shows the first statement and the second follows from repeating the argument with the dual basis \(\mathcal{B}^*\) and observing that \(\mathcal{B}^{**} = \mathcal{B}\text{.}\)
For the second statement \(\delta \in U_2^\circ\) then for any \(\mb{u} \in U_1 \) we have \(\delta (\mb{u} ) = 0 \) since \(U_1 \leq U_2\text{.}\) Thus \(\delta \in U_1^\circ\text{.}\)

Exercises 2.5.3 Exercises

1.

Implicit equations for planes in \(\mathbb{R}^3\) with coordinates \((x, y, z)\) are of the form
\begin{equation} a x + b y + c z = d \tag{2.5.2} \end{equation}
where \(a, b, c,\) and \(d\) are real numbers. If \(d = 0\) the plane is a vector subspace of \(\mathbb{R}^3\text{.}\) For each of the following planes, find a basis \(\mathcal{A}\) and give a parameterization defined by \([]_{\mathcal{A}} : \mathbb{R}^2 \to \mathbb{R}^3\text{.}\)
(a)
\begin{equation*} x - y + z = 0 \end{equation*}
(b)
\begin{equation*} y + 2z = 0 \end{equation*}

3.

For each matrix in Exercise 2.4.3.1, give a basis for the row space and a basis for the column space. What are the ranks of these matrices?

4.

Using Proposition 2.5.14, explain why you need only one linear equation to describe a plane in \(\mathbb{R}^3\) but two linear equations to describe a line.