Subsection 4.1.3 Definition of a Vector Space
Looking at
Example 4.1.8 and
Example 4.1.9 shows us that a set with two operations (which we will call addition and scalar multiplication) such that the set is closed under the two operations, and satisfies the same eight properties as
\(\R^n\text{.}\) We will refer to such a set with its two operations as a
vector space.
Definition 4.1.10.
Let \(V\) be a nonempty set. Suppose that elements of \(V\) can be added together and can be multiplied by scalars. The set \(V\text{,}\) together with these operations of addition and scalar multiplication, is called a vector space provided that
and the following properties hold for \(\mathbf{u}\text{,}\) \(\mathbf{v}\) and \(\mathbf{w}\) in \(V\) and scalars \(k\) and \(p\text{:}\)
Commutative Property of Addition: \(\mathbf{u}+\mathbf{v}=\mathbf{v}+\mathbf{u}.\)
Associative Property of Addition:\quad \((\mathbf{u}+\mathbf{v})+\mathbf{w}=\mathbf{u}+(\mathbf{v}+\mathbf{w}).\)
Existence of Additive Identity: \(\mathbf{u}+\mathbf{0}=\mathbf{u}.\)
Existence of Additive Inverse: \(\mathbf{u}+(-\mathbf{u})=\mathbf{0}.\)
Distributive Property over Vector Addition: \(k(\mathbf{u}+\mathbf{v})=k\mathbf{u}+k\mathbf{v}.\)
Distributive Property over Scalar Addition: \((k+p)\mathbf{u}=k\mathbf{u}+p\mathbf{u}.\)
Associative Property for Scalar Multiplication: \(k(p\mathbf{u})=(kp)\mathbf{u}.\)
Multiplication by \(1\text{:}\) \(1\mathbf{u}=\mathbf{u}.\)
We will refer to elements of \(V\) as vectors.
When scalars
\(k\) and
\(p\) in the above definition are required to be real numbers, as they usually are in this text, the vector space
\(V\) may be called a
real vector space or
vector space over the real numbers. In
Chapter 6, we will sometimes consider complex numbers and that means that our typical vector space will be
\(\mathbb{C}^n\) and our scalars will be complex numbers; no change in the definition above is needed, other than allowing scalars to be complex numbers, for a
vector space over the complex numbers.
We have already encountered two real vectors spaces in
Example 4.1.8 and
Example 4.1.9, namely
\(\mathbb{M}_{m,n}\) and
\(\mathbb{L}\text{.}\)
Sets of polynomials provide an important source of examples, so we review some basic facts. A polynomial with real coefficients in \(x\) is an expression
\begin{equation*}
p(x) = a_0 + a_1x + a_2x^2 + \cdots + a_nx^n
\end{equation*}
where \(a_{0}, a_{1}, a_{2}, \ldots, a_{n}\) are real numbers called the coefficients of the polynomial.
If all the coefficients are zero, the polynomial is called the zero polynomial and is denoted simply as \(0\text{.}\)
If \(p(x) \neq 0\text{,}\) the highest power of \(x\) with a nonzero coefficient is called the degree of \(p(x)\) and denoted as \(\mbox{deg}(p(x))\text{.}\) The degree of the zero polynomial is not defined. The coefficient of this highest power of \(x\) is called the leading coefficient of \(p(x)\text{.}\) Hence \(\mbox{deg}(3 + 5x) = 1\) and its leading coefficient is \(5\text{,}\) \(\mbox{deg}(1 + 3x - x^{2}) = 2\) and its leading coefficient is \(-1\text{,}\) and the constant polynomial \(p(x)=4\) had degree 0 and leading coefficient 4.
To simplify our calculations, for the polynomial \(p(x) = a_0 + a_1x + \cdots + a_nx^n\text{,}\) we define \(a_{n+1}, a_{n+2}, \ldots\) to all be zero. This means \(a_i\) will make sense for all non-negative values of \(i\text{.}\) For example, \(p(x)\) is the zero polynomial exactly when \(a_i=0\) for all non-negative integers \(i\text{.}\) Because \(p(x)\) is a polynomial, there are only finitely many \(i\) with \(a_i \neq 0\) and, when \(p(x)\) is not the zero polynomial, the largest such index \(i\) is the degree of \(p(x)\text{.}\)
Let \(\mathbb{P}\) denote the set of all polynomials and suppose that
\begin{align*}
p(x) \amp = a_0 + a_1x + a_2x^2 + \cdots + a_mx^m\\
q(x) \amp = b_0 + b_1x + b_2x^2 + \cdots + b_nx^n
\end{align*}
are two polynomials in \(\mathbb{P}\) (possibly of different degrees). Then \(p(x)\) and \(q(x)\) are called equal (written \(p(x) = q(x)\)) if and only if all the corresponding coefficients are equal--- that is, one has \(a_{0} = b_{0}\text{,}\) \(a_{1} = b_{1}\text{,}\) \(a_{2} = b_{2}\text{,}\) and so on. In particular, \(a_{0} + a_{1}x + \ldots + a_mx^m = 0\) means \(a_{0} = 0\text{,}\) \(a_{1} = 0\text{,}\) \(a_{2} = 0\text{,}\) \(\ldots\text{.}\)
The set \(\mathbb{P}\) has vector addition and scalar multiplication operations defined as follows: if \(p(x)\) and \(q(x)\) are as before and \(k\) is a real number,
\begin{align*}
p(x) + q(x) \amp = (a_0 + b_0) + (a_1 + b_1)x + (a_2 + b_2)x^2 + \cdots + (a_j+b_j) x^j\\
kp(x) \amp = ka_0 + (ka_1)x + (ka_2)x^2 + \cdots + (k a_m) x^m
\end{align*}
where \(j\) is the larger of \(m\) or \(n\text{.}\) Notice that the degree of \(p(x)+q(x)\) is at most the larger of the degrees of \(p(x)\) and \(q(x)\) and the degree of \(k(p(x))\) is exactly the degree of \(p(x)\text{,}\) provided \(k \ne 0\text{.}\)
This is a ton of terminology, so we need some examples to understand it.
Example 4.1.11.
\(\mathbb{P}\) is a vector space.
Answer.
It is easy to see that the sum of two polynomials is again a polynomial, and that a scalar multiple of a polynomial is a polynomial. Thus, \(\mathbb{P}\) is closed under addition and scalar multiplication. The other eight vector space properties can be verified (do one yourself), and we conclude that \(\mathbb{P}\) is a vector space.
Example 4.1.12.
Let \(Y\) be the set of all degree two polynomials in \(x\text{.}\) In other words,
\begin{equation*}
Y=\left \lbrace ax^2+bx+c : a \ne 0 \right \rbrace.
\end{equation*}
We claim that \(Y\) is not a vector space.
Answer.
Observe that \(Y\) is not closed under addition. To see this, let \(y_1 = 2x^2+3x+4\) and let \(y_2=-2x^2\text{.}\) Then \(y_1\) and \(y_2\) are both elements of \(Y\text{.}\) However, \(y_1+y_2 = 3x+4\) is not an element of \(Y\text{,}\) as it is only a degree one polynomial. We require the coefficient \(a\) of \(x^2\) to be nonzero for a polynomial to be in \(Y\text{,}\) and this is not the case for \(y_1+y_2\text{.}\) As an exercise, check the remaining vector space properties one-by-one to see which properties hold and which do not.
The set
\(Y\) in
Example 4.1.12 is not a vector space, but if we make a slight modification, we can make it into a vector space.
Example 4.1.13.
Let \(\mathbb{P}^2\) be the set of polynomials of degree two or less. In other words,
\begin{equation*}
\mathbb{P}^2=\left \lbrace ax^2+bx+c : a,b,c \in \mathbb{R} \right \rbrace.
\end{equation*}
Note that
\(\mathbb{P}^2\) contains the zero polynomial (let
\(a=b=c=0\)). Unlike set
\(Y\) in
Example 4.1.12,
\(\mathbb{P}^2\) is closed under polynomial addition and scalar multiplication. It is easy to verify that all vector space properties hold, so
\(\mathbb{P}^2\) is a vector space.
Example 4.1.14.
Let
\(n\) be a natural number. Define
\(\mathbb{P}^n\) to be the set of polynomials of degree
\(n\) or less than
\(n\text{,}\) then by reasoning similar to
Example 4.1.13,
\(\mathbb{P}^n\) is a vector space.
Subsection 4.1.4 Subspaces
Definition 4.1.15.
A nonempty subset \(U\) of a vector space \(V\) is called a subspace of \(V\text{,}\) provided that \(U\) is itself a vector space when given the same addition and scalar multiplication as \(V\text{.}\)
An example to showcase this is in order.
Example 4.1.16.
In
Example 4.1.13 we demonstrated that
\(\mathbb{P}^2\) is a vector space. From
Example 4.1.11 we know that
\(\mathbb{P}\) is a vector space. But
\(\mathbb{P}^2\) is a subset of
\(\mathbb{P}\text{,}\) and uses the same operations of polynomial addition and scalar multiplication. Therefore
\(\mathbb{P}^2\) is a subspace of
\(\mathbb{P}\text{.}\)
Checking all ten properties to verify that a subset of a vector space is a subspace can be cumbersome. Fortunately we have the following theorem.
Theorem 4.1.17. Subspace Test.
Let \(U\) be a nonempty subset of a vector space \(V\text{.}\) If \(U\) is closed under the operations of addition and scalar multiplication of \(V\text{,}\) then \(U\) is a subspace of \(V\text{.}\)
Proof.
To prove that closure is a sufficient condition for \(U\) to be a subspace, we will need to show that closure under addition and scalar multiplication of \(V\) guarantees that the remaining eight properties are satisfied automatically.
Observe that
Item 1,
Item 2,
Item 5,
Item 6,
Item 7 and
Item 8 hold for all elements of
\(V\text{.}\) Thus, these properties will hold for all elements of
\(U\text{.}\) We say that these properties are
inherited from
\(V\text{.}\)
To prove
Item 3 we need to show that
\(\mathbf{0}\text{,}\) which we know to be an element of
\(V\text{,}\) is contained in
\(U\text{.}\) Let
\(\mathbf{u}\) be an element of
\(U\) (recall that
\(U\) is nonempty). We will show that
\(0\mathbf{u}=\mathbf{0}\) in
\(V\text{.}\) Then, by closure under scalar multiplication, we will be able to conclude that
\(0\mathbf{u}=\mathbf{0}\) must be in
\(U\text{.}\)
\begin{equation*}
0\mathbf{u}=(0+0)\mathbf{u}=0\mathbf{u}+0\mathbf{u}.
\end{equation*}
Adding the additive inverse of \(0\mathbf{u}\) to both sides gives us
\begin{equation*}
0\mathbf{u}+(-0\mathbf{u})=(0\mathbf{u}+0\mathbf{u})+(-0\mathbf{u}).
\end{equation*}
\begin{equation*}
\mathbf{0}=0\mathbf{u}+(0\mathbf{u}+(-0\mathbf{u})).
\end{equation*}
\begin{equation*}
\mathbf{0}=0\mathbf{u}+\mathbf{0}=0\mathbf{u}.
\end{equation*}
Because \(U\) is closed under scalar multiplication \(0\mathbf{u}=\mathbf{0}\) is in \(U\text{.}\) We know that every element of \(U\text{,}\) being an element of \(V\text{,}\) has an additive inverse in \(V\text{.}\) We need to show that the additive inverse of every element of \(U\) is contained in \(U\text{.}\) Let \(\mathbf{u}\) be any element of \(U\text{.}\) We will show that \((-1)\mathbf{u}\) is the additive inverse of \(\mathbf{u}\text{.}\) Then by closure, \((-1)\mathbf{u}\) will have to be contained in \(U\text{.}\) To show that \((-1)\mathbf{u}\) is the additive inverse of \(\mathbf{u}\text{,}\) we must show that \(\mathbf{u}+(-1)\mathbf{u}=\mathbf{0}\text{.}\) We compute:
\begin{equation*}
\mathbf{u}+(-1)\mathbf{u}=1\mathbf{u}+(-1)\mathbf{u}=(1+(-1))\mathbf{u}=0\mathbf{u}=\mathbf{0}.
\end{equation*}
Thus \((-1)\mathbf{u}\) is the additive inverse of \(\mathbf{u}\text{.}\) By closure, \((-1)\mathbf{u}\) is in \(U\text{.}\)
Let us apply the theorem in a couple of examples.
Example 4.1.18.
Let \(V\) be the set of vectors in \(\R^3\) on the \(y-\)axis. Then \(V\) is a subspace of \(\R^3\text{.}\)
Answer.
To verify this, note that any vector in \(V\) is of the form
\begin{equation*}
\begin{bmatrix}0\\y\\0 \end{bmatrix}\text{.}
\end{equation*}
If we multiply such a vector by a scalar \(c\text{,}\) we get a vector of the form
\begin{equation*}
\begin{bmatrix}0\\cy\\0 \end{bmatrix},
\end{equation*}
which is clearly still in \(V\text{.}\) This proves \(V\) is closed under scalar multiplication. Next, note that
\begin{equation*}
\begin{bmatrix}0\\y_1\\0 \end{bmatrix} + \begin{bmatrix}0\\y_2\\0 \end{bmatrix}= \begin{bmatrix}0\\y_1 + y_2\\0\end{bmatrix},
\end{equation*}
so \(V\) is closed under addition.
Example 4.1.19.
Let \(A\) be a fixed matrix in \(\mathbb{M}_{n,n}\text{.}\) Show that the set \(C_A\) of all \(n\times n\) matrices that commute with \(A\) under matrix multiplication is a subspace of \(\mathbb{M}_{n,n}\text{.}\)
Answer.
The set \(C_A\) consists of all \(n\times n\) matrices \(X\) such that \(AX=XA\text{.}\) First, observe that \(C_A\) is not empty because \(I_n\) is an element. Now we need to show that \(C_A\) is closed under matrix addition and scalar multiplication. Suppose that \(X_1\) and \(X_{2}\) lie in \(C_A\text{.}\) Then \(AX_1 = X_1A\) and \(AX_{2} = X_{2}A\text{.}\) Then
\begin{equation*}
A(X_1 + X_2) = AX_1 + AX_2 = X_1A + X_2A + (X_1 + X_2)A.
\end{equation*}
Therefore \((X_1+X_2)\) commutes with \(A\text{.}\) Thus \((X_1+X_2)\) is in \(C_A\text{.}\) We conclude that \(C_A\) is closed under matrix addition. Now suppose \(X\) is in \(C_A\text{.}\) Let \(k\) be a scalar, then
\begin{equation*}
A(kX)= k(AX) = k(XA) = (kX)A.
\end{equation*}
Therefore \((kX)\) commutes with \(A\text{.}\) We conclude that \((kX)\) is in \(C_A\text{,}\) and \(C_A\) is closed under scalar multiplication. Hence \(C_A\) is a subspace of \(\mathbb{M}_{n,n}\text{.}\)
To get used to the new terminology, let us look at an example in the context of polynomials.
Example 4.1.21.
Consider the set \(U\) of all polynomials in \(\mathbb{P}\) that have \(3\) as a root:
\begin{equation*}
U = \lbrace p(x) \in \mathbb{P} : p(3) = 0 \rbrace.
\end{equation*}
Show that \(U\) is a subspace of \(\mathbb{P}\text{.}\)
Answer.
Observe that \(U\) is not empty because \(r(x)=x-3\) is an element of \(U\text{.}\) Suppose \(p(x)\) and \(q(x)\) lie in \(U\text{.}\) Then \(p(3) = 0\) and \(q(3) = 0\text{.}\) We have
\begin{equation*}
(p + q)(x) = p(x) + q(x)
\end{equation*}
for all \(x\text{,}\) so
\begin{equation*}
(p + q)(3) = p(3) + q(3) = 0 + 0 = 0,
\end{equation*}
and \(U\) is closed under addition. The verification that \(U\) is closed under scalar multiplication is similar.
The following important results provides us with a quick way to determine that some subsets are not subspaces.
Theorem 4.1.22.
If \(W\) is a subspace of a vector space \(V\text{,}\) then the zero vector \(\mathbf{0}\) is in \(W\text{.}\)
Proof.
Since a subspace is nonempty, there is some vector \(\mathbf{w}\) in \(W\text{.}\) Then \(0 \mathbf{w} = \mathbf{0}\) is in \(W\) because \(W\) is closed under scalar multiplication.
Consider lines in
\(\R^n\text{.}\) Applying
Theorem 4.1.22 shows that the only lines in
\(\R^n\) that are subspaces are those that pass through the origin. The same holds true for planes and hyperplanes. For example, the plane
\(z=3\) in
\(\R^3\) is not a subspace of
\(\R^3\text{.}\) It turns out that any plane containing the origin is a subspace.
Be warned that containing the zero vector is a necessary condition but not a sufficient condition to be a subspace. That is, if a subset does not contain the zero vector, then it is not a subspace. But if a subset does contain the zero vector, it may or may not be a subspace. For example, the subset of \(\R^2\) given by \(\{ (x,y) : y=x^2 \}\text{,}\) the graph of the parabola \(y=x^2\text{,}\) contains the zero vector of \(\R^2\text{,}\) but is not a subspace since it is not closed under addition. (Don’t belive me? Just check!)
Theorem 4.1.23.
If \(W\) is a subspace of a vector space \(V\text{,}\) then for any vector \(\mathbf{w} \in W\text{,}\) the additive inverse, \(-\mathbf{w}\text{,}\) is also in \(W\text{.}\)
The proof is similar to what was done for the previous theorem and is left as an exercise.