Letβs recall what we know about vectors in \(\R^2\text{.}\) Writing \(\vv = \langle x,y\rangle\) for the vector pointing from \((0,0)\) to \((x,y)\text{,}\) we define:
We can then observe a number of properties enjoyed by these operations. In your first course, you may have observed some of these properties geometrically, using the βtip-to-tailβ rule for vector addition, as shown in FigureΒ 1.1.1
Vector addition is commutative. That is, for any vectors \(\vv = \langle a, b\rangle\) and \(\ww = \langle c,d\rangle\text{,}\) we have \(\vv+\ww = \ww+\vv\text{.}\)
Vector addition is associative. That is, for any vectors \(\uu = \langle a, b\rangle, \vv = \langle c, d\rangle\) and \(\ww = \langle p, q\rangle\text{,}\) we have
This tells us that placement of parentheses doesnβt matter, which is essential for extending addition (which is defined as an operation on two vectors) to sums of three or more vectors.
Vector addition has an identity element. This is a vector that has no effect when added to another vector, or in other words, the zero vector. Again, it inherits its property from the behaviour of the real number 0.
Every vector has an inverse with respect to addition, or, in other words, a negative. Given a vector \(\vv = \langle a, b\rangle\text{,}\) the vector \(-\vv = \langle -a, -b\rangle\) satisfies
Scalar multiplication is compatible with addition in two different ways. First, it is distributive over vector addition: for any scalar \(k\) and vectors \(\vv=\langle a, b\rangle, \ww = \langle c,d\rangle\text{,}\) we have \(k(\vv+\ww)=k\vv+k\ww\text{.}\)
Second, scalar multiplication is also distributive with respect to scalar addition: for any scalars \(c\) and \(d\) and vector \(\vv\text{,}\) we have \((c+d)\vv=c\vv+d\vv\text{.}\)
Finally, there is a βnormalizationβ result for scalar multiplication. For any vector \(\vv\text{,}\) we have \(1\vv = \vv\text{.}\) That is, the real number \(1\) acts as an identity element with respect to scalar multiplication. (You can check this one yourself.)
You might be wondering why we bother to list the last property above. Itβs true, but why do we need it? One reason comes from basic algebra, and solving equations. Suppose we have the equation \(c\vv = \ww\text{,}\) where \(c\) is some nonzero scalar, and we want to solve for \(\vv\text{.}\) Very early in our algebra careers, we learn that to solve, we βdivide by \(c\)β.
Division doesnβt quite make sense in this context, but it certainly does make sense to multiply both sides by \(1/c\text{,}\) the multiplicative inverse of \(c\text{.}\) We then have \(\left(\frac1c\right)(c\vv)=\left(\frac1c\right)\ww\text{,}\) and since scalar multiplication is associative, \(\left(\frac1c \cdot c\right)\vv = \left(\frac1c\right)\ww\text{.}\) We know that \(\frac1c \cdot c = 1\text{,}\) so this boils down to \(1\vv=(1/c)\ww\text{.}\) It appears that weβve solved the equation, but only if we know that \(1\vv = \vv\text{.}\)
For an example where this fails, take our vectors as above, but redefine the scalar multiplication as \(c\langle a, b\rangle = \langle ca, 0\rangle\text{.}\) The distributive and associative properties for scalar multiplication will still hold, but the normalization property fails. Algebra becomes very strange with this version of scalar multiplication. In particular, we can no longer conclude that if \(2\vv=2\ww\text{,}\) then \(\vv=\ww\text{!}\)
Given an example of vectors \(\vv\) and \(\ww\) such that \(2\vv=2\ww\text{,}\) but \(\vv\neq \ww\text{,}\) if scalar multiplication is defined as above.
In a first course in linear algebra, these algebraic properties of vector addition and scalar multiplication are presented as a theorem. (After all, we have just demonstrated the truth of these results.) A second course in linear algebra (and in particular, abstract linear algebra), begins by taking that theorem and turning it into a definition. We will then do some exploration, to see if we can come up with some other examples that fit the definition; the significance of this is that we can expect the algebra in these examples to behave in essentially the same way as the vectors weβre familiar with.
A real vector space (or vector space over \(\R\)) is a nonempty set \(V\text{,}\) whose objects are called vectors, equipped with two operations:
Addition, which is a map from \(V\times V\) to \(V\) that associates each ordered pair of vectors \((\vv,\ww)\) to a vector \(\vv+\ww\text{,}\) called the sum of \(\vv\) and \(\ww\text{.}\)
Scalar multiplication, which is a map from \(\R\times V\) to \(V\) that associates each real number \(c\) and vector \(\vv\) to a vector \(c\vv\text{.}\)
Note that a zero vector must exist in every vector space. This simple observation is a key component of many proofs and counterexamples in linear algebra. In general, we may define a vector space whose scalars belong to a field\(\mathbb{F}\text{.}\) A field is a set of objects whose algebraic properties are modelled after those of the real numbers.
The axioms for a field are not all that different than those for a vector space. The main difference is that in a field, multiplication is defined between elements of the field (and produces another element of the field), while scalar multiplication combines elements of two different sets.
Note how the axioms for multiplication in a field mirror the addition axioms much more closely than in a vector space. The only difference is the fact that there is one element without a multiplicative inverse; namely, the zero element.
While it is possible to study linear algebra over finite fields (like the integers modulo a prime number) we will only consider two fields: the real numbers \(\R\text{,}\) and the complex numbers \(\C\text{.}\)
Before we move on, letβs look at one example involving finite fields. Let \(\mathbb{Z}_n=\{0,1,2,\ldots, n-1\}\text{,}\) with addition and multiplication defined modulo \(n\text{.}\) (For example, \(3+5=1\) in \(\mathbb{Z}_7\text{,}\) since \(8\equiv 1 \pmod{7}\text{.}\))
A vector space whose scalars are complex numbers will be called a complex vector space. While many students are initially intimidated by the complex numbers, most results in linear algebra work exactly the same over \(\C\) as they do over \(\R\text{.}\) And where the results differ, things are usually easier with complex numbers, owing in part to the fact that all complex polynomials can be completely factored.
The following are examples of vector spaces. We leave verification of axioms as an exercise. (Verification will follow a process very similar to the discussion at the beginning of this section.)
The set \(\R^n\) of \(n\)-tuples \((x_1,x_2,\ldots, x_n)\) of real numbers, where we define
We will also often use \(\R^n\) to refer to the vector space of \(1\times n\) column matrices \(\bbm x_1\\x_2\\\vdots\\x_n\ebm\text{,}\) where addition and scalar multiplication are defined as for matrices (and the same as the above, with the only difference being the way in which we choose to write our vectors). If the distinction between \(n\)-tuples and column matrices is ever important, it will be made clear.
Addition and scalar multiplication are defined in the same way as \(\R^n\text{;}\) the only difference is that elements of \(\R^\infty\) contain infinitely many entries.
This is the same as the addition and scalar multiplication we get for functions in general, using the βpointwise evaluationβ definition: for polynomials \(p\) and \(q\) and a scalar \(c\text{,}\) we have \((p+q)(x)=p(x)+q(x)\) and \((cp)(x)=c\cdot p(x)\text{.}\)
Notice that although this feels like a very different example, the vector space \(P_n(\R)\) is in fact very similar to \(\R^n\) (or rather, \(\R^{n+1}\text{,}\) to be precise). If we associate the polynomial \(a_0+a_1x+\cdots + a_nx^n\) with the vector \(\langle a_0,a_1,\ldots, a_n\rangle\text{,}\) the addition and scalar multiplication for either space behaves in exactly the same way. We will make this observation precise in SectionΒ 2.3.
The set \(P(\R)\) of all polynomials of any degree. The algebra works the same as it does in \(P_n(\R)\text{,}\) but there is an important difference: in both \(P_n(\R)\) and \(\R^n\text{,}\) every element in the set can be generated by setting values for a finite collection of coefficients. (In \(P_n(\R)\text{,}\) every polynomial \(a_0+a_1x+\cdots =a_nx^n\) can be obtained by choosing values for the \(n+1\) coefficients \(a_0,a_1\ldots, a_n\text{.}\)) But if we remove the restriction on the degree of our polynomials, there is then no limit on the number of coefficients we might need. (Even if any individual polynomial has a finite number of coefficients!)
The set \(F[a,b]\) of all functions \(f:[a,b]\to \R\text{,}\) where we define \((f+g)(x)=f(x)+g(x)\) and \((cf)(x)=c(f(x))\text{.}\) The zero function is the function satisfying \(0(x)=0\) for all \(x\in [a,b]\text{,}\) and the negative of a function \(f\) is given by \((-f)(x)=-f(x)\) for all \(x\in [a,b]\text{.}\)
Note that while the vector space \(P(\R)\) has an infinite nature that \(P_n(\R)\) does not, the vector space \(F[a,b]\) is somehow more infinite! Using the language of SectionΒ 1.7, we can say that \(P_n(\R)\) is finite dimensional, while \(P(\R)\) and \(F[a,b]\) are infinite dimensional. In a more advanced course, one might make a further distinction: the dimension of \(P(\R)\) is countably infinite, while the dimension of \(F[a,b]\) is uncountable.
How should we define addition and scalar multiplication? Since the usual addition and scalar multiplication wont work, letβs denote addition by \(x\oplus y\text{,}\) for \(x,y\in V\text{,}\) and scalar multiplication by \(c\odot x\text{,}\) for \(c\in\R\) and \(x\in V\text{.}\)
Note: you can format any math in your answers using LaTeX, by putting a $ before and after the math. For example, \(x\oplus y\) is $x\oplus y$, and \(x\odot y\) is $x\odot y$.
Remember that an inverse \(-x\) must satisfy \(x\oplus (-x)=e\text{,}\) where \(e\) is the identity element. What is \(e\text{,}\) and how is βadditionβ defined?
True or false: the set of all polynomials with real number coefficients and degree less than or equal to three is a vector space, using the usual polynomial addition and scalar multiplication.
True or false: the set of all polynomials with real number coefficients and degree greater than or equal to three, together with the zero polynomial, is a vector space, using the usual polynomial addition and scalar multiplication.
True or false: the set of all vectors \(\vv = \langle a, b\rangle\) of unit length (that is, such that \(\sqrt{a^2+b^2}=1\)) is a vector space with respect to the usual addition and scalar multiplication in \(\R^2\text{.}\)
Let \(V={\mathbb R}\text{.}\) For \(u,v \in V\) and \(a\in{\mathbb R}\) define vector addition by \(u \boxplus v := u+v + 3\) and scalar multiplication by \(a \boxdot u := au + 3a - 3\text{.}\) It can be shown that \((V,\boxplus,\boxdot)\) is a vector space over the scalar field \(\mathbb R\text{.}\) Find the following:
Let \(V=(-5,\infty)\text{.}\) For \(u,v \in V\) and \(a\in{\mathbb R}\) define vector addition by \(u \boxplus v := uv + 5(u+v)+20\) and scalar multiplication by \(a \boxdot u := (u + 5)^a - 5\text{.}\) It can be shown that \((V,\boxplus,\boxdot)\) is a vector space over the scalar field \(\mathbb R\text{.}\) Find the following: