Published on

Linear Algebra Done Right Ch2: Finite-Dimensional Vector Spaces

Authors
  • avatar
    Name
    走歪的工程師James
    Twitter

In the previous chapter, we understood what an abstract vector space is. Linear algebra focuses mainly on finite-dimensional vector spaces. But what does 'finite dimension' mean? How do you define dimension? We know that (2,1)(2, 1) is two-dimensional, (2,1,2)(2, 1, -2) is three-dimensional, and RnR^n is n-dimensional; these are all finite dimensional. But if a vector is not in the form of a list, how do we define its dimension? What does "dimension" actually mean?

Actually, the dimension of a vector space, is defined by "the number of vectors required to span the space".

As an example, R2\mathbf{R}^2 can be spanned by (1,0),(0,1)(1, 0), (0, 1). this means that, any point on the R2\mathbf{R}^2 plane, can be decomposed into (1,0)(1, 0) multiplied by a scalar, plus (0,1)(0, 1) multiplied by a scalar.

a1(1,0)+a2(0,1),a1,a2Ra_1(1, 0) + a_2(0, 1), a_1, a_2 \in \mathbf{R}

Using the set notation, we have

R2={a1(1,0)+a2(0,1)a1,a2R}\mathbf{R}^2 = \{a_1(1, 0) + a_2(0, 1) \mid a_1, a_2 \in \mathbf{R} \}

On top of that, (1,0),(0,1)(1, 0), (0, 1) are not the only vectors that work. Pick any 2 vectors in R2\mathbf{R}^2, and you will find that as long as they are not scalar multiples of each other(in the same direction), then they span R2\mathbf{R}^2. So the choice of vectors don't seem to matter.

R2={a1(1,1)+a2(0,1)a1,a2R}\mathbf{R}^2 = \{a_1(1, 1) + a_2(0, -1) \mid a_1, a_2 \in \mathbf{R} \}

However, if we reduce the list to 1 vector, it cannot span R2\mathbf{R}^2 anymore, whichever vector you pick. So it seems like the 2D plan requires at least 2 vectors to span, while the 3D space requires at least 3 vectors to span. This gives a basis for an intuition of the definition of dimension.

A 2-dimensional space can be spanned by 2 vectors.

A 3-dimensional space can be spanned by 3 vectors.

That's pretty much what dimension is! We've explained it in a loose, intuitive manner.

Now, let us explore this further and see how linear algebra formally defines and arrives at the definition of "dimension"!

2.A Span and Linear Independence

When we talk about list of vectors, we often omit the surrounding parentheses. For example:

  • When we say (1,0),(0,1)(1, 0), (0, 1) spans R2\mathbf{R}^2, we actually mean the list ((1,0),(0,1))((1, 0), (0, 1)) spans R2\mathbf{R}^2.
  • (4,1,6),(9,5,7)(4, 1, 6), (9, 5, 7) is a list of length two of vectors in R3\mathbf{R}^3.

So let's get this definition out of the way.

definition 2.1:list of vectors

We will usually write lists of vectors without surrounding parentheses.

Back to the main topic. We previously touch upon the notion of "the sum of scalar multiples of vectors". We call this a linear combination. Here we formally define it.

definition 2.2: linear combination

A linear combination of a list v1,,vmv_1, \ldots, v_m of vectors in VV is a vector of the form

a1v1++amvma_1 v_1+\cdots+a_m v_m

where a1,,amFa_1, \ldots, a_m \in \mathbf{F}.

For example: (17,4,2)(17, −4, 2) is a linear combination of (2,1,3),(1,2,4)(2, 1, −3), (1, −2, 4), because

(17,4,2)=6(2,1,3)+5(1,2,4)(17, −4, 2) = 6(2, 1, −3) + 5(1, −2, 4)

However, (17,4,5)(17, −4, 5) is not a linear combination of (2,1,3),(1,2,4)(2, 1, −3), (1, −2, 4), because no a1,a2Fa_1, a_2 \in \mathbf{F} exist such that

(17,4,5)=a1(2,1,3)+a2(1,2,4)(17, −4, 5) = a_1(2, 1, −3) + a_2(1, −2, 4)

Now we formally define "span".

definition 2.4: span

The set of all linear combinations of a list of vectors v1,,vmv_1, \ldots, v_m in VV is called the span of v1,,vmv_1, \ldots, v_m, denoted by span(v1,,vm)\operatorname{span}\left(v_1, \ldots, v_m\right). In other words,

span(v1,,vm)={a1v1++amvm:a1,,amF}\operatorname{span}\left(v_1, \ldots, v_m\right)=\left\{a_1 v_1+\cdots+a_m v_m: a_1, \ldots, a_m \in \mathbf{F}\right\}

The span of the empty list ( ) is defined to be {0}\{0\}.

For example,

  • (17,4,2)span((2,1,3),(1,2,4))(17, −4, 2) \in \operatorname{span}\left((2, 1, −3), (1, −2, 4)\right)
  • (17,4,5)span((2,1,3),(1,2,4))(17, −4, 5) \notin \operatorname{span}\left((2, 1, −3), (1, −2, 4)\right)

definition 2.9: finite-dimensional vector spaces

A vector space is called finite-dimensional if some list of vectors in it spans the space.

For instance, for any positive integer nn, Fn\mathbf{F}^n is a finite-dimensional vector space. because:

  • Any vector in Fn\mathbf{F}^n can be written as (x1,,xn)(x_1, \ldots, x_n)
  • (x1,,xn)=x1(1,0,,0)+x2(0,1,0..0)++xn(0,,0,1)\left(x_1, \ldots, x_n\right)=x_1(1,0, \ldots, 0)+x_2(0,1,0 \ldots . .0)+\cdots+x_n(0, \ldots, 0,1)
  • Hence (x1,,xn)span((1,0,,0),(0,1,0,,0),,(0,,0,1))\left(x_1, \ldots, x_n\right) \in \operatorname{span}((1,0, \ldots, 0),(0,1,0, \ldots, 0), \ldots,(0, \ldots, 0,1))

Polynomials, which we've seen a lot an high school math, can actually also be seen as vectors! Is it finite dimensional or not though? Before we explore that, let's first introduce some notations.

definition 2.10: polynomial,𝒫(𝐅)

  • A function p:FFp: \mathbf{F} \rightarrow \mathbf{F} is called a polynomial with coefficients in F\mathbf{F} if there exist a0,,amFa_0, \ldots, a_m \in \mathbf{F} such that
p(z)=a0+a1z+a2z2++amzmp(z)=a_0+a_1 z+a_2 z^2+\cdots+a_m z^m

for all zFz \in \mathbf{F}.

  • P(F)\mathcal{P}(\mathbf{F}) is the set of all polynomials with coefficients in F\mathbf{F}.

We introduce 2 concepts here:

  • First, we look at polynomials from a new perspective: as a function that maps a number to a number(FF\mathbf{F} \rightarrow \mathbf{F})
  • Second, we define the notation P(F)\mathcal{P}(\mathbf{F}) to denote the set of all polynomials

As two examples:

  • x37x+5P(R)x^3 - 7x + 5 \in \mathcal{P}(\mathbf{R})
  • (3+i)z2iz+5P(C)(3+i)z^2 - iz + 5 \in \mathcal{P}(\mathbf{C})

Having defined the set P(F)\mathcal{P}(\mathbf{F}), you might have guess what I'm going to say next. That's right! We claim that P(F)\mathcal{P}(\mathbf{F}) is a vector space. You are encouraged to verify all the conditions for a vector space. Refer back to the last chapter if needed.

Remember that we mentioned that FS\mathbf{F}^{\mathbf{S}} is a vector space in the last chapter? This means FF\mathbf{F}^{\mathbf{F}}, being a special case, a subspace of it.

Now we define Pm(F)\mathcal{P}_m(\mathbf{F}), the set of all polynomials with degree no more than mm

definition 2.11:degree of a polynomial, deg 𝑝

  • A polynomial pP(F)p \in \mathcal{P}(\mathbf{F}) is said to have degree mm if there exist scalars a0,a1,,amFa_0, a_1, \ldots, a_m \in \mathbf{F} with am0a_m \neq 0 such that for every zFz \in \mathbf{F}, we have
p(z)=a0+a1z++amzmp(z)=a_0+a_1 z+\cdots+a_m z^m
  • The polynomial that is identically 0 is said to have degree -\infty.
  • The degree of a polynomial pp is denoted by degp\operatorname{deg} p.

notation 2.12:notation: 𝒫𝑚(𝐅)

For mm a nonnegative integer, Pm(F)\mathcal{P}_m(\mathbf{F}) denotes the set of all polynomials with coefficients in F\mathbf{F} and degree at most mm.

It's easy to see Pm(F)=span(1,z,,zm)\mathcal{P}_m(\mathbf{F}) = \operatorname{span}(1, z, \ldots, z^m). In other words, a finite list of vectors spans Pm(F)\mathcal{P}_m(\mathbf{F}), implying Pm(F)\mathcal{P}_m(\mathbf{F}) is a finite-dimensional vector space.

definition 2.13:definition: infinite-dimensional vector space

A vector space is called infinite-dimensional if it is not finite-dimensional.

This definition may seem a bit repetitive. But let's think through it to see what it means.

  • Finite-dimensional is defined to mean "can be spanned by a finite list of vectors"
  • Infinite-dimensional is defined to mean "not finite-dimensional"
  • Therefore, a vector space that cannot be spanned by a finite list of vectors is infinite-dimensional

As an example, P(R)\mathcal{P}(\mathbf{R}) is infinite-dimensional. Why?

  • Think of any list of members P(R)\mathcal{P}(\mathbf{R})
  • Let mm denote the hight degree of the polynomials in this list
  • Then every polynomial in the span of this list has degree at most mm
  • Thus zm+1z^{m+1} is not in said span
  • Hence no list spans P(R)\mathcal{P}(\mathbf{R})
  • Thus P(R)\mathcal{P}(\mathbf{R}) is infinite dimensional

Linear Independence

We have understood that the 2D plane can be spanned by 2 vectors. But 2 vectors on the same line cannot span the 2D plane. This distinction is formalized by the concept of "linear independence".

definition 2.15: linearly independent

  • A list v1,,vmv_1, \ldots, v_m of vectors in VV is called linearly independent if the only choice of a1,,amFa_1, \ldots, a_m \in \mathbf{F} that makes a1v1++amvm=0a_1 v_1+\cdots+a_m v_m=0 is a1==am=0a_1=\cdots=a_m=0.
  • The empty list ( ) is also declared to be linearly independent.

A list of vectors (v1,,vm)(v_1, \ldots, v_m) is linearly independent if and only if

  • 00 has only 1 way to be written as a linear combination(taking all coefficients to be 0)
  • Any vector in span(v1,,vm)\operatorname{span}(v_1, \ldots, v_m) has only one representation as a linear combination of v1,,vmv_1, \ldots, v_m

Pause and think about why this is the case. Visualize it with the 2D or 3D cases. What does it look like when your list of vectors is linear independent, and when it is linear dependent?

Now we look at a few examples:

  1. (1,0,0,0),(0,1,0,0),(0,0,1,0)(1,0,0,0), (0,1,0,0), (0,0,1,0) is linearly independent in F4\mathbf{F}^4. To see why, assume a1,a2,a3Fa_1, a_2, a_3 \in \mathbf{F}. Then we have a1(1,0,0,0)+a2(0,1,0,0)+a3(0,0,1,0)=(0,0,0,0)a_1(1,0,0,0)+a_2(0,1,0,0)+a_3(0,0,1,0)=(0,0,0,0) Hence (a1,a2,a3,0)=(0,0,0,0)(a_1,a_2,a_3,0)=(0,0,0,0) Thus a1=a2=a3=0a_1=a_2=a_3=0. So (1,0,0,0),(0,1,0,0),(0,0,1,0)(1,0,0,0), (0,1,0,0), (0,0,1,0) is linearly independent in F4\mathbf{F}^4.
  2. 1,z,,zm1, z, \ldots, z^m is linearly independent in P(R)\mathcal{P}(\mathbf{R})
  3. A list of one vector is linearly independent if and only if it is not the 00 vector
  4. A list of 2 vectors is linearly independent     \iff neither of them is a scalar multiple of the other

If a list of vectors is not linearly independent, they are linearly dependent.

definition 2.17: linearly dependent

  • A list of vectors in VV is called linearly dependent if it is not linearly independent.
  • In other words, a list v1,,vmv_1, \ldots, v_m of vectors in VV is linearly dependent if there exist a1,,amFa_1, \ldots, a_m \in \mathbf{F}, not all 0 , such that a1v1++amvm=0a_1 v_1+\cdots+a_m v_m=0.

Here are a few examples. Make sure to understand each of them:

  • (2,3,1),(1,1,2),(7,3,8)(2, 3, 1), (1, −1, 2), (7, 3, 8) is linearly dependent in F3\mathbf{F}^3
  • In VV, if a list of vectors has a vector that is a linear combination of the other vectors, then this list is linearly dependent.(proof: Writing that one vector in the list as equal to a linear combination of the other vectors, move that vector to the other side of the equation to get a linear combination of 00 that is not all 0 coefficients)
  • Every list of vectors in VV containing the 00 vector is linearly dependent.(a special case of the previous bullet point)

The next lemma is a useful tool that we will use a lot for many proofs we will see later.

definition 2.19:linear dependence lemma

Suppose v1,,vmv_1, \ldots, v_m is a linearly dependent list in VV. Then there exists k{1,2,,m}k \in\{1,2, \ldots, m\} such that

vkspan(v1,,vk1)v_k \in \operatorname{span}\left(v_1, \ldots, v_{k-1}\right)

Furthermore, if kk satisfies the condition above and the kth k^{\text {th }} term is removed from v1,,vmv_1, \ldots, v_m, then the span of the remaining list equals span(v1,,vm)\operatorname{span}\left(v_1, \ldots, v_m\right).

Proof: This is a straightforward proof. We want to prove that:

  1. There exists a number kk such that the kk-th vector is a linear combination of the vectors preceding it, i.e. vkspan(v1,,vk1)v_k \in \operatorname{span}(v_1, \ldots, v_{k-1})
  2. Removing vkv_k from the list does not change the span

We first prove the first bullet point.

  • v1,,vmv_1, \ldots, v_m are linearly dependent, which means there exist a1,,amFa_1, \ldots, a_m \in F, not all zero, such that a1v1++amvm=0a_1 v_1 + \ldots + a_m v_m = 0.
  • Suppose aka_k is the last non-zero coefficient among a1,,ama_1, \ldots, a_m, then we can rearrange to get vk=a1akv1ak1akvk1v_k = -\frac{a_1}{a_k} v_1 - \cdots - \frac{a_{k-1}}{a_k} v_{k-1}.
  • This implies vkspan(v1,,vk1)v_k \in \operatorname{span}\left(v_1, \ldots, v_{k-1}\right), as desired.

Now we come to the second point.

  • Suppose uu is a vector in the span, that is, uspan(v1,,vm)u \in \operatorname{span}(v_1, \ldots, v_m).
  • uu can be written as a linear combination of these vectors: u=c1v1++cmvmu = c_1 v_1 + \cdots + c_m v_m.
  • From the first point, we know that vkv_k can be replaced by a linear combination of v1,,vk1v_1, \ldots, v_{k-1}.
  • Since uu is any vector, it means that any vector in the span can be formed without needing vkv_k, thus proving the point.

We've completed the proof. But how should we interpret this theorem? I essentially says that in a linearly dependent list of vectors, we can find one vector that is a linear combination of the previous elements. Thus even if the throw out this "redundant" vector, the span of the list does not become smaller.

Conversely, when you adjoin to a list a vector already in the span of of the list, the span does not grow bigger. Only linearly independent vectors contribute to the span.

Let us look at a few examples:

example 2.21

(1,2,3),(6,5,4),(15,16,17),(8,9,7)(1, 2, 3), (6, 5, 4), (15, 16, 17), (8, 9, 7) are linearly dependent in R3\mathbf{R}^3. To see why,

  • If we take k=1k=1, the first vector must be 0. But since (1,2,3)(1, 2, 3) is not the zero vector, we cannot take k=1k=1.
  • If we take k=2k=2, the second vector must be a multiple of the first vector, but there does not exist cRc \in \mathbf{R} such that (6,5,4)=c(1,2,3)(6,5,4)=c(1,2,3), so we cannot take k=2k=2.
  • Taking k=3k=3, we can find (15,16,17)=3(1,2,3)+2(6,5,4)(15, 16, 17) = 3(1, 2, 3) + 2(6, 5, 4), thus this set of vectors is linearly dependent.

With this lemma proven, we now prepared to prove a key result.

theorem 2.22:length of linearly independent list <= length of spanning list

In a finite-dimensional vector space, the length of every linearly independent list of vectors is less than or equal to the length of every spanning list of vectors.

Proof: We first assume:

  • u1,,umu_1, \ldots, u_m are linearly dependent in VV
  • w1,,wnw_1, \ldots, w_n spans VV

We prove mnm \leq n by the following process. Note that in each step we add one of the uu's and remove one of the ww's.

Step 1:

  • Let BB be the list w1,,wnw_1, \ldots, w_n that spans VV.
  • Insert u1u_1 at the beginning of BB, which will produce a linearly dependent list u1,w,,wlinearly dependent\overbrace{u_1, w, \ldots, w}^{\text{linearly dependent}}
  • According to 2.19 (the linear dependence lemma), one of the vectors in the list can be expressed as a combination of the preceding vectors, and thus can be removed. This "redundant vector" will not be u1u_1, because u1u_1 is not in {0}\{0\} (where {0}\{0\} is the span of the empty list).
  • In other words, one of the ww's can be removed without affecting the span of the list.

Step kk (k=2,,mk=2, \ldots, m):

  • The list obtained from step k1k-1 spans VV.
  • Therefore, we can insert uku_k into the list, forming a linearly dependent list (the subscripts of the ww's are omitted here). u1,,uk,w,,wlinearly dependent\overbrace{u_1, \ldots, u_k, w, \ldots, w}^{\text{linearly dependent}}
  • Because this list is linearly dependent, we can remove one of the vectors without affecting its span.
  • It is important to note that the first part of this list (composed of the uu's) is linearly independent and thus will not be removed. Since we need to find a vector that can be composed of the preceding vectors, the one to be removed must be from the ww part. u1,,uklinearly independent,w,,wlinearly dependent\overbrace{\underbrace{u_1, \ldots, u_k}_{\mathclap{\text{linearly independent}}}, w, \ldots, w}^{\text{linearly dependent}}

Repeat process above, which adds a uu and removes a ww in each step, and we will find that all of the uu's have been moved to the other list, implying there are more ww's than uu's, as we expect.

This proof may be confusing at first glance, so here is some additional explanation:

  1. Every time we insert a uu, we get a linearly dependent list
  2. By lemma 2.19, there MUST be a vector that we can move without changing the span
  3. This vector CANNOT be a uu, because all of the uu's are linearly independent. Thus it must be one of the ww's
  4. This implies there are more ww's than uu's

In plain English, this theorem states that a linearly independent list of vectors is always shorter than a spanning list, the explanation for which is that once a list already spans the entire space, any new vector adjoined to it must already be in the span, thus forming a linearly dependent list.

This theorem implies an interesting result: in some cases we know without any calculation whether a list of vector is linearly dependent or not.

example 2.23

  • The list (1,0,0),(0,1,0),(0,0,1)(1, 0, 0), (0, 1, 0), (0, 0, 1) of length 3 spans R3\mathbf{R}^3.
  • This also means that in R3\mathbf{R}^3, no list of length greater than 3 can be linearly independent!
  • For example, the list (1,2,3),(4,5,8),(9,6,7),(3,2,8)(1, 2, 3), (4, 5, 8), (9, 6, 7), (−3, 2, 8) has a length of four, and therefore cannot be linearly independent!

example 2.24

  • The list (1,0,0,0),(0,1,0,0),(0,0,1,0),(0,0,0,1)(1, 0, 0, 0), (0, 1, 0, 0), (0, 0, 1, 0), (0, 0, 0, 1) is of length 4 and is linearly independent.
  • Therefore, any list that spans R4\mathbf{R}^4 must have a length greater than 4, meaning no list of fewer than 4 vectors can span R4\mathbf{R}^4.
  • For example, the list (1,2,3,5),(4,5,8,3),(9,6,7,1)(1, 2, 3, -5), (4, 5, 8, 3), (9, 6, 7, -1) has a length of 3, and therefore cannot span R4\mathbf{R}^4.

Hand-picked exercises(from Linear Algebra Done Right, Ch2.A

  1. (a) Show that a list of length one in a vector space is linearly independent if and only if the vector in the list is not 0.

    (b) Show that a list of length two in a vector space is linearly independent if and only if neither of the vectors in the list are not scalar multiples of the other.

  1. (a) Show that if we think of C\mathbf{C} as a vector space over R\mathbf{R}, then the list 1+i,1i1+i, 1-i is linearly independent.

    (b) Show that if we think of C\mathbf{C} as a vector space over C\mathbf{C}, then the list 1+i,1i1+i, 1-i is linearly dependent.

  1. Explain why there does not exist a list of six polynomials that is linearly independent in P4(F)\mathcal{P}_4(\mathbf{F}).

2.B Bases

definition 2.26: Bases

A basis of VV is a list of vectors in VV that is linearly independent and spans VV.

Let us look at some examples:

  1. The list (1,0,,0),(0,1,0,,0),,(0,,0,1)(1,0, \ldots, 0),(0,1,0, \ldots, 0), \ldots,(0, \ldots, 0,1) is a basis of Fn\mathbf{F}^n, called the standard basis of Fn\mathbf{F}^n.
  2. The list (1,2),(3,5)(1,2),(3,5) is a basis of F2\mathbf{F}^2. Note that this list has length two, which is the same as the length of the standard basis of F2\mathbf{F}^2. In the next section, we will see that this is not a coincidence.
  3. The list (1,2,4),(7,5,6)(1,2,-4),(7,-5,6) is linearly independent in F3\mathbf{F}^3 but is not a basis of F3F^3 because it does not span F3F^3.
  4. The list (1,2),(3,5),(4,13)(1,2),(3,5),(4,13) spans F2\mathbf{F}^2 but is not a basis of F2\mathbf{F}^2 because it is not linearly independent.
  5. The list (1,1,0),(0,0,1)(1,1,0),(0,0,1) is a basis of {(x,x,y)F3:x,yF}\left\{(x, x, y) \in \mathbf{F}^3: x, y \in \mathbf{F}\right\}.
  6. The list (1,1,0),(1,0,1)(1,-1,0),(1,0,-1) is a basis of
{(x,y,z)F3:x+y+z=0}.\left\{(x, y, z) \in \mathbf{F}^3: x+y+z=0\right\} .
  1. The list 1,z,,zm1, z, \ldots, z^m is a basis of Pm(F)\mathcal{P}_m(\mathbf{F}), called the standard basis of Pm(F)\mathcal{P}_m(\mathbf{F}).

Make sure you work through each example in your brain!

definition 2.28: criterion for basis

A list v1,,vnv_1, \ldots, v_n of vectors in VV is a basis of VV if and only if every vVv \in V can be written uniquely in the form

v=a1v1++anvnv=a_1 v_1+\cdots+a_n v_n

where a1,,anFa_1, \ldots, a_n \in \mathbf{F}.

This is a bi-directional proof. We first prove the right arrow direction:

(    )(\implies)

  • Assume v1,,vnv_1, \ldots, v_n is a basis for VV
  • Let vv be a vector in VV
  • Because v1,,vnv_1, \ldots, v_n spans VV there exist a1,,anFa_1, \ldots, a_n \in \mathbf{F} such that 2.29 holds
  • To show that the representation in 2.29 is unique, suppose c1,,cnc_1, \ldots, c_n are scalars such that we also have v=c1v1++cnvnv=c_1 v_1+\cdots+c_n v_n
  • Subtracting the last equation from 2.29, we get 0=(a1c1)v1++(ancn)vn.0=\left(a_1-c_1\right) v_1+\cdots+\left(a_n-c_n\right) v_n .
  • This implies that each akcka_k-c_k equals 0 (because v1,,vnv_1, \ldots, v_n is linearly independent).
  • Hence a1=c1,,an=cna_1=c_1, \ldots, a_n=c_n. We have the desired uniqueness, completing the proof in one direction.

Now, the other direction: (    )(\impliedby)

  • suppose every vVv \in V can be written uniquely in the form given by 2.29.
  • This implies that the list v1,,vnv_1, \ldots, v_n spans VV.
  • To show that v1,,vnv_1, \ldots, v_n is linearly independent, suppose a1,,anFa_1, \ldots, a_n \in \mathbf{F} are such that 0=a1v1++anvn0=a_1 v_1+\cdots+a_n v_n
  • The uniqueness of the representation 2.29 (taking v=0v=0 ) now implies that a1==an=0a_1=\cdots=a_n=0
  • Thus v1,,vnv_1, \ldots, v_n is linearly independent and hence is a basis of VV.

theorem 2.30:every spanning list reduces to a basis

Every spanning list in a vector space can be reduced to a basis of the vector space.

Proof:

We first assume v1,,vnv_1, \ldots, v_n spans VV. Applying the following algorithm gives a basis.

Step 1:

  • If v1=0v_1=0, then delete v1v_1 from BB. If v10v_1 \neq 0, then leave BB unchanged.

Step kk

  • If vkv_k is in span(v1,,vk1)\operatorname{span}\left(v_1, \ldots, v_{k-1}\right), then delete vkv_k from the list BB.
  • If vkv_k is not in span(v1,,vk1)\operatorname{span}\left(v_1, \ldots, v_{k-1}\right), then leave BB unchanged.

The process stops after kk steps. We denote the remaining list by BB. BB spans VV because we only removed vectors that were linear combinations of previous vectors. This algorithm ensures BB does not have any "redundant" vector. Hence by 2.19, BB is linearly independent, forming a basis for VV.

theorem 2.31: basis of finite-dimensional space

Every finite-dimensional vector space has a basis.

  • By definition, a finite-dimensional vector space has a spanning list.
  • The previous result tells us that each spanning list can be reduced to a basis

theorem 2.32: linearly independent lists extend to bases

Every linearly independent list of vectors in a finite-dimensional vector space can be extended to a basis of the vector space.

Proof:

  • Suppose u1,,umu_1, \ldots, u_m is linearly independent in a finite-dimensional vector space VV
  • Let w1,,wnw_1, \ldots, w_n be a list of vectors in VV that spans VV
  • Thus the list u1,,um,w1,,wnu_1, \ldots, u_m, w_1, \ldots, w_n spans VV.
  • Applying the procedure of the proof of 2.30 to reduce this list to a basis of VV
  • produces a basis consisting of the vectors u1,,umu_1, \ldots, u_m and some of the ww 's (none of the uu 's get deleted in this procedure because u1,,umu_1, \ldots, u_m is linearly independent).

As an example in F3\mathbf{F}^3:

  • Suppose we start with the linearly independent list (2,3,4),(9,6,8)(2,3,4),(9,6,8).
  • Take w1,w2,w3w_1, w_2, w_3 to be the standard basis of F3\mathbf{F}^3
  • Applying the procedure in the proof above produces the list (2,3,4),(9,6,8),(0,1,0)(2,3,4),(9,6,8),(0,1,0) which is a basis of F3\mathbf{F}^3.

Up until now, we have learned the new concept of a "basis". A basis combines the concepts of "linear independence" and "spanning lists" introduced earlier. A basis of VV

  1. spans VV
  2. is linearly independent

Think of it as "the shortest list of vectors that spans a vector space". It has just the right length—not too many to lose linear independence, and not too few to fail to span the space.

Having understood the basis, in the next section, we can define the concept of "dimension"!

2.B Hand-picked exercises(From Linear Algebra Done Right, Ch2.B

  1. (a) Let UU be the subspace of R5\mathbf{R}^5 defined by

    U={(x1,x2,x3,x4,x5)R5:x1=3x2 and x3=7x4}.U=\left\{\left(x_1, x_2, x_3, x_4, x_5\right) \in \mathbf{R}^5: x_1=3 x_2 \text { and } x_3=7 x_4\right\} .

    Find a basis of UU.

    (b) Extend the basis in (a) to a basis of R5\mathbf{R}^5.

    (c) Find a subspace WW of R5\mathbf{R}^5 such that R5=UW\mathbf{R}^5=U \oplus W.

  1. Prove or give a counterexample: If p0,p1,p2,p3p_0, p_1, p_2, p_3 is a list in P3(F)\mathcal{P}_3(\mathbf{F}) such that none of the polynomials p0,p1,p2,p3p_0, p_1, p_2, p_3 has degree 2 , then p0,p1,p2,p3p_0, p_1, p_2, p_3 is not a basis of P3(F)\mathcal{P}_3(\mathbf{F}).
  1. Suppose v1,v2,v3,v4v_1, v_2, v_3, v_4 is a basis of VV. Prove that v1+v2,v2+v3,v3+v4,v4v_1+v_2, v_2+v_3, v_3+v_4, v_4 is also a basis of VV.
  1. Suppose UU and WW are subspaces of VV such that V=UWV=U \oplus W. Suppose also that u1,,umu_1, \ldots, u_m is a basis of UU and w1,,wnw_1, \ldots, w_n is a basis of WW. Prove that u1,,um,w1,,wnu_1, \ldots, u_m, w_1, \ldots, w_n is a basis of VV.

2.C Dimension

Now we finally come to define "dimension", which we mentioned in the beginning. Intuitively, the definition should be such that

  • F2\mathbf{F}^2 has dimension 2
  • F3\mathbf{F}^3 has dimension 3
  • Fn\mathbf{F}^n has dimension nn

The concept of a basis may lead you to the intuition of

"Maybe we can define dimension as the length of a basis?"

However, a finite-dimensional vector space in general has many different bases. Our attempted definition only makes sense if all bases for a vector space have the same length. Otherwise, we wouldn't have a consistent dimension for a vector space.

Fortunately, the is not an issue. All bases do have the same length, as we now show.

theorem 2.34: every basis has the same length

Any two bases of a finite-dimensional vector space have the same length.

Recall 2.22:

In a finite-dimensional vector space, the length of every linearly independent list of vectors is less than or equal to the length of every spanning list of vectors.

We now show every basis has the same length using 2.22:

  • Suppose VV is finite-dimensional. Let B1B_1 and B2B_2 be two bases of VV.
  • Then B1B_1 is linearly independent in VV and B2B_2 spans VV, so the length of 𝐵1 is at most the length of B2B_2 (by 2.22).
  • Interchanging the roles of B1B_1 and B2B_2, we also see that the length of B2B_2 is at most the length of 𝐵1.
  • Thus the length of B1B_1 equals the length of B2B_2, as desired

Now that we know that any two bases of a finite-dimensional vector space have the same length, we are ready to define the dimension of such spaces.

definition 2.35: dimension、dim V

  • The dimension of a finite-dimensional vector space is the length of any basis of the vector space.
  • The dimension of a finite-dimensional vector space VV is denoted by dimV\operatorname{dim} V.
  • dimFn=n\operatorname{dim} \mathbf{F}^n=n because the standard basis of Fn\mathbf{F}^n has length nn.
  • dimPm(F)=m+1\operatorname{dim} \mathcal{P}_m(\mathbf{F})=m+1 because the standard basis 1,z,,zm1, z, \ldots, z^m of Pm(F)\mathcal{P}_m(\mathbf{F}) has length m+1m+1.
  • If U={(x,x,y)F3:x,yF}U=\left\{(x, x, y) \in \mathbf{F}^3: x, y \in \mathbf{F}\right\}, then dimU=2\operatorname{dim} U=2 because (1,1,0),(0,0,1)(1,1,0),(0,0,1) is a basis of UU.
  • If U={(x,y,z)F3:x+y+z=0}U=\left\{(x, y, z) \in \mathbf{F}^3: x+y+z=0\right\}, then dimU=2\operatorname{dim} U=2 because the list (1,1,0),(1,0,1)(1,-1,0),(1,0,-1) is a basis of UU.

By definition, a basis of a vector space VV must

  1. be linearly independent
  2. span VV

Now we show that if a list of vector has the right length, then one of the properties being true implies the other holds as well!

For example:

  • If there are three linearly independent vectors in R3\mathbf{R}^3, then they will definitely span VV, so they form a basis.
  • If there are three vectors that span R3\mathbf{R}^3, then they are definitely linearly independent in VV, so they form basis.

Now we prove this intuition is true.

definition 2.38: linearly independent list of the right length is a basis

Suppose VV is finite-dimensional. Then every linearly independent list of vectors in VV of length dimV\operatorname{dim} V is a basis of VV.

Proof

  • Suppose dimV=n\operatorname{dim} V=n and v1,,vnv_1, \ldots, v_n is linearly independent in VV.
  • The list v1,,vnv_1, \ldots, v_n can be extended to a basis of VV (by 2.32).
  • However, every basis of VV has length nn, so in this case the extension is the trivial one, meaning that no elements are adjoined to v1,,vnv_1, \ldots, v_n.
  • Thus v1,,vnv_1, \ldots, v_n is a basis of VV, as desired.

Now we prove that a spanning list of the right length must be linearly independent, and thus is a basis.

definition 2.42: spanning list of the right length is a basis

Suppose VV is finite-dimensional. Then every spanning list of vectors in VV of length dimV\operatorname{dim} V is a basis of VV.

Proof

  • Suppose dimV=n\operatorname{dim} V=n and v1,,vnv_1, \ldots, v_n spans VV.
  • The list v1,,vnv_1, \ldots, v_n can be reduced to a basis of VV (by 2.30).
  • However, every basis of VV has length nn, so in this case the reduction is the trivial one, meaning that no elements are deleted from v1,,vnv_1, \ldots, v_n.
  • Thus v1,,vnv_1, \ldots, v_n is a basis of VV, as desired.

The next result gives a formula for the dimension of the sum of two subspaces of a finite-dimensional vector space. The resembles the formula for the number of elements in the union of two sets:

# elements in first set+# elements in second set# elements in intersection\text{\# elements in first set} + \text{\# elements in second set} - \text{\# elements in intersection}

Writing it with formal notations:

#(S1S2)=#S1+#S2#(S1S2)\begin{aligned} \#\left(S_1 \cup S_2\right) & =\# S_1+\# S_2-\#\left(S_1 \cap S_2\right)\end{aligned}

We now prove this formula.

definition 2.43: definition of a sum

If V1V_1 and V2V_2 are subspaces of a finite-dimensional vector space, then

dim(V1+V2)=dimV1+dimV2dim(V1V2).\operatorname{dim}\left(V_1+V_2\right)=\operatorname{dim} V_1+\operatorname{dim} V_2-\operatorname{dim}\left(V_1 \cap V_2\right) .

Proof:

  • Let v1,,vmv_1, \ldots, v_m be a basis of V1V2V_1 \cap V_2; thus dim(V1V2)=m\operatorname{dim}\left(V_1 \cap V_2\right)=m.
  • Because v1,,vmv_1, \ldots, v_m is a basis of V1V2V_1 \cap V_2, it is linearly independent in V1V_1.
  • Hence this list can be extended to a basis v1,,vm,u1,,ujv_1, \ldots, v_m, u_1, \ldots, u_j of V1V_1 (by 2.32). Thus dimV1=m+j\operatorname{dim} V_1=m+j.
  • Also extend v1,,vmv_1, \ldots, v_m to a basis v1,,vm,w1,,wkv_1, \ldots, v_m, w_1, \ldots, w_k of V2V_2; thus dimV2=m+k\operatorname{dim} V_2=m+k.

We will show that

v1,,vm,u1,,uj,w1,,wk(2.44)v_1, \ldots, v_m, u_1, \ldots, u_j, w_1, \ldots, w_k \tag{2.44}

is a basis of V1+V2V_1+V_2. This will complete the proof, because then we will have

dim(V1+V2)=m+j+k=(m+j)+(m+k)m=dimV1+dimV2dim(V1V2).\begin{aligned} \operatorname{dim}\left(V_1+V_2\right) & =m+j+k \\ & =(m+j)+(m+k)-m \\ & =\operatorname{dim} V_1+\operatorname{dim} V_2-\operatorname{dim}\left(V_1 \cap V_2\right) . \end{aligned}

To get started, we know that

  • The list 2.44 is contained in V1V2V_1 \cup V_2 and thus is contained in V1+V2V_1+V_2.
  • The span of this list contains V1V_1 and contains V2V_2 and hence is equal to V1+V2V_1+V_2.
  • Thus to show that 2.44 is a basis of V1+V2V_1+V_2 we only need to show that it is linearly independent.

To prove that 2.44 is linearly independent, suppose

a1v1++amvm+b1u1++bjuj+c1w1++ckwk=0a_1 v_1+\cdots+a_m v_m+b_1 u_1+\cdots+b_j u_j+c_1 w_1+\cdots+c_k w_k=0

where all the aa 's, bb 's, and cc 's are scalars.

We need to prove that all the aa 's, bb 's, and cc 's equal 0 . The equation above can be rewritten as

c1w1++ckwk=a1v1amvmb1u1bjujV1(2.45)c_1 w_1+\cdots+c_k w_k=\underbrace{-a_1 v_1-\cdots-a_m v_m-b_1 u_1-\cdots-b_j u_j}_{\in V_1} \tag{2.45}

which shows that c1w1++ckwkV1c_1 w_1+\cdots+c_k w_k \in V_1.

All the ww 's are in V2V_2, so this implies that c1w1++ckwkV1V2c_1 w_1+\cdots+c_k w_k \in V_1 \cap V_2.

Because v1,,vmv_1, \ldots, v_m is a basis of V1V2V_1 \cap V_2, we have

c1w1++ckwk=d1v1++dmvmc_1 w_1+\cdots+c_k w_k=d_1 v_1+\cdots+d_m v_m

for some scalars d1,,dmd_1, \ldots, d_m. Moving all the terms to the same side, we have

c1w1++ckwkd1v1dmvm=0c_1 w_1+\cdots+c_k w_k - d_1 v_1 - \cdots - d_m v_m=0

But v1,,vm,w1,,wkv_1, \ldots, v_m, w_1, \ldots, w_k is linearly independent, so this implies that all the cc 's (and dd 's) equal 0 . Thus 2.45 becomes the equation

a1v1++amvm+b1u1++bjuj=0a_1 v_1+\cdots+a_m v_m+b_1 u_1+\cdots+b_j u_j=0

Because the list v1,,vm,u1,,ujv_1, \ldots, v_m, u_1, \ldots, u_j is linearly independent, this equation implies that all the aa 's and bb 's are 0 , completing the proof.

2.C Hand-picked exercises(From Linear Algebra Done Right, Ch2.C

  1. Show that the subspaces of R2\mathbf{R}^2 are precisely {0}\{0\}, all lines in R2\mathbf{R}^2 containing the origin, and R2\mathbf{R}^2.
  1. (a) Let U={pP4(F):p(6)=0}U=\left\{p \in \mathcal{P}_4(\mathbf{F}): p(6)=0\right\}. Find a basis of UU.

    (b) Extend the basis in (a) to a basis of P4(F)\mathcal{P}_4(\mathbf{F}).

    (c) Find a subspace WW of P4(F)\mathcal{P}_4(\mathbf{F}) such that P4(F)=UW\mathcal{P}_4(\mathbf{F})=U \oplus W.

  1. Suppose v1,,vmv_1, \ldots, v_m is linearly independent in VV and wVw \in V. Prove that
dimspan(v1+w,,vm+w)m1\operatorname{dim} \operatorname{span}\left(v_1+w, \ldots, v_m+w\right) \geq m-1
  1. Suppose UU and WW are both five-dimensional subspaces of R9\mathbf{R}^9. Prove that UW{0}U \cap W \neq\{0\}.
  1. Explain why you might guess, motivated by analogy with the formula for the number of elements in the union of three finite sets, that if V1,V2,V3V_1, V_2, V_3 are subspaces of a finite-dimensional vector space, then dim(V1+V2+V3)=dimV1+dimV2+dimV3dim(V1V2)dim(V1V3)dim(V2V3)+dim(V1V2V3)\begin{aligned} \operatorname{dim}\left(V_1+V_2\right. & \left.+V_3\right) \\ = & \operatorname{dim} V_1+\operatorname{dim} V_2+\operatorname{dim} V_3 \\ & -\operatorname{dim}\left(V_1 \cap V_2\right)-\operatorname{dim}\left(V_1 \cap V_3\right)-\operatorname{dim}\left(V_2 \cap V_3\right) \\ & +\operatorname{dim}\left(V_1 \cap V_2 \cap V_3\right) \end{aligned} Then either prove the formula above or give a counterexample.

Summary

We started this chapter with an intuition for dimension and then formally define and prove our way up to its formal definition.

  • First, we defined "linear combinations" and "spans"
  • Second, we defined "linear dependence"
  • Then, we combined both concepts and defined "bases"
  • Finally, we proved that all bases have the same length to show our definition makes sense

References