86
Chapter 3 Direct Sums, Ane Maps, The Dual Space, Duality 3.1 Direct Products, Sums, and Direct Sums There are some useful ways of forming new vector spaces from older ones. Definition 3.1. Given p 2 vector spaces E 1 ,...,E p , the product F = E 1 ··· E p can be made into a vector space by defining addition and scalar multiplication as follows: (u 1 ,...,u p )+(v 1 ,...,v p )=(u 1 + v 1 ,...,u p + v p ) λ(u 1 ,...,u p )=(λu 1 ,..., λu p ), for all u i ,v i 2 E i and all λ 2 R. With the above addition and multiplication, the vector space F = E 1 ··· E p is called the direct product of the vector spaces E 1 ,...,E p . 151

Chapter 3 Direct Sums, Ane Maps, The Dual Space, Dualitycis515/cis515-15-sl1-c.pdf · 3.1. DIRECT PRODUCTS, SUMS, AND DIRECT SUMS 153 It is clear that this map is linear, and so its

  • Upload
    others

  • View
    17

  • Download
    0

Embed Size (px)

Citation preview

Chapter 3

Direct Sums, A�ne Maps, The DualSpace, Duality

3.1 Direct Products, Sums, and Direct Sums

There are some useful ways of forming new vector spacesfrom older ones.

Definition 3.1. Given p � 2 vector spaces E1, . . . , Ep,the product F = E1⇥ · · ·⇥Ep can be made into a vectorspace by defining addition and scalar multiplication asfollows:

(u1, . . . , up) + (v1, . . . , vp) = (u1 + v1, . . . , up + vp)

�(u1, . . . , up) = (�u1, . . . , �up),

for all ui, vi 2 Ei and all � 2 R.

With the above addition and multiplication, the vectorspace F = E1 ⇥ · · · ⇥ Ep is called the direct product ofthe vector spaces E1, . . . , Ep.

151

152 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

The projection maps pri : E1 ⇥ · · · ⇥ Ep ! Ei given by

pri(u1, . . . , up) = ui

are clearly linear.

Similarly, the maps ini : Ei ! E1 ⇥ · · · ⇥ Ep given by

ini(ui) = (0, . . . , 0, ui, 0, . . . , 0)

are injective and linear.

It can be shown (using bases) that

dim(E1 ⇥ · · · ⇥ Ep) = dim(E1) + · · · + dim(Ep).

Let us now consider a vector space E and p subspacesU1, . . . , Up of E.

We have a map

a : U1 ⇥ · · · ⇥ Up ! E

given bya(u1, . . . , up) = u1 + · · · + up,

with ui 2 Ui for i = 1, . . . , p.

3.1. DIRECT PRODUCTS, SUMS, AND DIRECT SUMS 153

It is clear that this map is linear, and so its image is asubspace of E denoted by

U1 + · · · + Up

and called the sum of the subspaces U1, . . . , Up.

By definition,

U1 + · · · + Up = {u1 + · · · + up | ui 2 Ui, 1 i p},

and it is immediately verified that U1 + · · · + Up is thesmallest subspace of E containing U1, . . . , Up.

If the map a is injective, then Ker a = {( 0, . . . , 0| {z }p

)},

which means that if ui 2 Ui for i = 1, . . . , p and if

u1 + · · · + up = 0

then u1 = 0, . . . , up = 0.

In this case, every u 2 U1 + · · · + Up has a unique ex-pression as a sum

u = u1 + · · · + up,

with ui 2 Ui, for i = 1, . . . , p.

154 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

It is also clear that for any p nonzero vectors ui 2 Ui,u1, . . . , up are linearly independent.

Definition 3.2. For any vector space E and any p � 2subspaces U1, . . . , Up of E, if the map a defined above isinjective, then the sum U1 + · · · + Up is called a directsum and it is denoted by

U1 � · · · � Up.

The space E is the direct sum of the subspaces Ui if

E = U1 � · · · � Up.

Observe that when the map a is injective, then it is alinear isomorphism between U1 ⇥ · · · ⇥ Up andU1 � · · · � Up.

The di↵erence is that U1 ⇥ · · · ⇥ Up is defined even ifthe spaces Ui are not assumed to be subspaces of somecommon space.

There are natural injections from each Ui to E denotedby ini : Ui ! E.

3.1. DIRECT PRODUCTS, SUMS, AND DIRECT SUMS 155

Now, if p = 2, it is easy to determine the kernel of themap a : U1 ⇥ U2 ! E. We have

a(u1, u2) = u1 + u2 = 0 i↵ u1 = �u2, u1 2 U1, u2 2 U2,

which implies that

Ker a = {(u, �u) | u 2 U1 \ U2}.

Now, U1 \ U2 is a subspace of E and the linear mapu 7! (u, �u) is clearly an isomorphism between U1 \ U2

and Ker a, so Ker a is isomorphic to U1 \ U2.

As a consequence, we get the following result:

Proposition 3.1. Given any vector space E and anytwo subspaces U1 and U2, the sum U1 + U2 is a directsum i↵ U1 \ U2 = (0).

156 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Recall that an n ⇥ n matrix A 2 Mn is symmetric ifA> = A, skew -symmetric if A> = �A. It is clear that

S(n) = {A 2 Mn | A> = A}Skew(n) = {A 2 Mn | A> = �A}

are subspaces of Mn, and that S(n) \ Skew(n) = (0).

Observe that for any matrix A 2 Mn, the matrix H(A) =(A + A>)/2 is symmetric and the matrixS(A) = (A � A>)/2 is skew-symmetric. Since

A = H(A) + S(A) =A + A>

2+

A � A>

2,

we have the direct sum

Mn = S(n) � Skew(n).

3.1. DIRECT PRODUCTS, SUMS, AND DIRECT SUMS 157

Proposition 3.2. Given any vector space E and anyp � 2 subspaces U1, . . . , Up, the following propertiesare equivalent:

(1) The sum U1 + · · · + Up is a direct sum.

(2) We have

Ui \✓ pX

j=1,j 6=i

Uj

◆= (0), i = 1, . . . , p.

(3) We have

Ui \✓ i�1X

j=1

Uj

◆= (0), i = 2, . . . , p.

The isomorphism U1 ⇥ · · · ⇥ Up ⇡ U1 � · · · � Up implies

Proposition 3.3. If E is any vector space, for any(finite-dimensional) subspaces U1, . . ., Up of E, wehave

dim(U1 � · · · � Up) = dim(U1) + · · · + dim(Up).

158 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

If E is a direct sum

E = U1 � · · · � Up,

since every u 2 E can be written in a unique way as

u = u1 + · · · + up

for some ui 2 Ui for i = 1 . . . , p, we can define the maps⇡i : E ! Ui, called projections , by

⇡i(u) = ⇡i(u1 + · · · + up) = ui.

It is easy to check that these maps are linear and satisfythe following properties:

⇡j � ⇡i =

(⇡i if i = j

0 if i 6= j,

⇡1 + · · · + ⇡p = idE.

3.1. DIRECT PRODUCTS, SUMS, AND DIRECT SUMS 159

For example, in the case of the direct sum

Mn = S(n) � Skew(n),

the projection onto S(n) is given by

⇡1(A) = H(A) =A + A>

2,

and the projection onto Skew(n) is given by

⇡2(A) = S(A) =A � A>

2.

Clearly, H(A) + S(A) = A, H(H(A)) = H(A),S(S(A)) = S(A), and H(S(A)) = S(H(A)) = 0.

A function f such that f�f = f is said to be idempotent .Thus, the projections ⇡i are idempotent.

Conversely, the following proposition can be shown:

160 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Proposition 3.4. Let E be a vector space. For anyp � 2 linear maps fi : E ! E, if

fj � fi =

(fi if i = j

0 if i 6= j,

f1 + · · · + fp = idE,

then if we let Ui = fi(E), we have a direct sum

E = U1 � · · · � Up.

We also have the following proposition characterizing idem-potent linear maps whose proof is also left as an exercise.

Proposition 3.5. For every vector space E, iff : E ! E is an idempotent linear map, i.e., f�f = f ,then we have a direct sum

E = Ker f � Im f,

so that f is the projection onto its image Im f .

We are now ready to prove a very crucial result relatingthe rank and the dimension of the kernel of a linear map.

3.1. DIRECT PRODUCTS, SUMS, AND DIRECT SUMS 161

Theorem 3.6. Let f : E ! F be a linear map. Forany choice of a basis (f1, . . . , fr) of Im f , let (u1, . . . , ur)be any vectors in E such that fi = f (ui), for i =1, . . . , r. If s : Im f ! E is the unique linear map de-fined by s(fi) = ui, for i = 1, . . . , r, then s is injective,f � s = id, and we have a direct sum

E = Ker f � Im s

as illustrated by the following diagram:

Ker f // E = Ker f � Im sf//

Im f ✓ F.soo

As a consequence,

dim(E) = dim(Ker f )+dim(Im f ) = dim(Ker f )+rk(f ).

Remark: The dimension dim(Ker f ) of the kernel of alinear map f is often called the nullity of f .

We now derive some important results using Theorem 3.6.

162 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Proposition 3.7. Given a vector space E, if U andV are any two subspaces of E, then

dim(U) + dim(V ) = dim(U + V ) + dim(U \ V ),

an equation known as Grassmann’s relation.

The Grassmann relation can be very useful to figure outwhether two subspace have a nontrivial intersection inspaces of dimension > 3.

For example, it is easy to see that in R5, there are sub-spaces U and V with dim(U) = 3 and dim(V ) = 2 suchthat U \ V = (0).

However, we can show that if dim(U) = 3 and dim(V ) =3, then dim(U \ V ) � 1.

As another consequence of Proposition 3.7, if U and Vare two hyperplanes in a vector space of dimension n, sothat dim(U) = n � 1 and dim(V ) = n � 1, we have

dim(U \ V ) � n � 2,

and so, if U 6= V , then

dim(U \ V ) = n � 2.

3.1. DIRECT PRODUCTS, SUMS, AND DIRECT SUMS 163

Proposition 3.8. If U1, . . . , Up are any subspaces ofa finite dimensional vector space E, then

dim(U1 + · · · + Up) dim(U1) + · · · + dim(Up),

and

dim(U1 + · · · + Up) = dim(U1) + · · · + dim(Up)

i↵ the Uis form a direct sum U1 � · · · � Up.

Another important corollary of Theorem 3.6 is the fol-lowing result:

Proposition 3.9. Let E and F be two vector spaceswith the same finite dimension dim(E) = dim(F ) =n. For every linear map f : E ! F , the followingproperties are equivalent:

(a) f is bijective.

(b) f is surjective.

(c) f is injective.

(d) Ker f = (0).

164 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

One should be warned that Proposition 3.9 fails in infinitedimension.

We also have the following basic proposition about injec-tive or surjective linear maps.

Proposition 3.10. Let E and F be vector spaces, andlet f : E ! F be a linear map. If f : E ! F isinjective, then there is a surjective linear map r : F !E called a retraction, such that r�f = idE. If f : E !F is surjective, then there is an injective linear maps : F ! E called a section, such that f � s = idF .

The notion of rank of a linear map or of a matrix impor-tant, both theoretically and practically, since it is the keyto the solvability of linear equations.

Proposition 3.11. Given a linear map f : E ! F ,the following properties hold:

(i) rk(f ) + dim(Ker f ) = dim(E).

(ii) rk(f ) min(dim(E), dim(F )).

3.1. DIRECT PRODUCTS, SUMS, AND DIRECT SUMS 165

The rank of a matrix is defined as follows.

Definition 3.3. Given a m ⇥ n-matrix A = (ai j), therank rk(A) of the matrix A is the maximum number oflinearly independent columns of A (viewed as vectors inRm).

In view of Proposition 1.4, the rank of a matrix A isthe dimension of the subspace of Rm generated by thecolumns of A.

Let E and F be two vector spaces, and let (u1, . . . , un) bea basis of E, and (v1, . . . , vm) a basis of F . Let f : E !F be a linear map, and let M(f ) be its matrix w.r.t. thebases (u1, . . . , un) and (v1, . . . , vm).

166 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Since the rank rk(f ) of f is the dimension of Im f , whichis generated by (f (u1), . . . , f (un)), the rank of f is themaximum number of linearly independent vectors in(f (u1), . . . , f (un)), which is equal to the number of lin-early independent columns of M(f ), since F and Rm areisomorphic.

Thus, we have rk(f ) = rk(M(f )), for every matrix rep-resenting f .

We will see later, using duality, that the rank of a ma-trix A is also equal to the maximal number of linearlyindependent rows of A.

3.1. DIRECT PRODUCTS, SUMS, AND DIRECT SUMS 167

Figure 3.1: How did Newton start a business

168 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

3.2 A�ne Maps

We showed in Section 1.5 that every linear map f mustsend the zero vector to the zero vector, that is,

f (0) = 0.

Yet, for any fixed nonzero vector u 2 E (where E is anyvector space), the function tu given by

tu(x) = x + u, for all x 2 E

shows up in pratice (for example, in robotics).

Functions of this type are called translations . They arenot linear for u 6= 0, since tu(0) = 0 + u = u.

More generally, functions combining linear maps and trans-lations occur naturally in many applications (robotics,computer vision, etc.), so it is necessary to understandsome basic properties of these functions.

3.2. AFFINE MAPS 169

For this, the notion of a�ne combination turns out toplay a key role.

Recall from Section 1.5 that for any vector space E, givenany family (ui)i2I of vectors ui 2 E, an a�ne combina-tion of the family (ui)i2I is an expression of the form

X

i2I

�iui withX

i2I

�i = 1,

where (�i)i2I is a family of scalars.

A linear combination places no restriction on the scalarsinvolved, but an a�ne combination is a linear combina-tion, with the restriction that the scalars �i must addup to 1. Nevertheless, a linear combination can alwaysbe viewed as an a�ne combination, using 0 with the co-e�cient 1 �

Pi2I �i.

A�ne combinations are also called barycentric combina-tions .

Although this is not obvious at first glance, the condi-tion that the scalars �i add up to 1 ensures that a�necombinations are preserved under translations.

170 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

To make this precise, consider functions f : E ! F ,where E and F are two vector spaces, such that thereis some linear map h : E ! F and some fixed vectorb 2 F (a translation vector ), such that

f (x) = h(x) + b, for all x 2 E.

The map f given by✓

x1

x2

◆7!

✓8/5 �6/53/10 2/5

◆✓x1

x2

◆+

✓11

is an example of the composition of a linear map with atranslation.

We claim that functions of this type preserve a�ne com-binations.

3.2. AFFINE MAPS 171

Proposition 3.12. For any two vector spaces E andF , given any function f : E ! F defined such that

f (x) = h(x) + b, for all x 2 E,

where h : E ! F is a linear map and b is some fixedvector in F , for every a�ne combination

Pi2I �iui

(withP

i2I �i = 1), we have

f

✓X

i2I

�iui

◆=X

i2I

�if (ui).

In other words, f preserves a�ne combinations.

Surprisingly, the converse of Proposition 3.12 also holds.

172 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Proposition 3.13. For any two vector spaces E andF , let f : E ! F be any function that preserves a�necombinations, i.e., for every a�ne combinationP

i2I �iui (withP

i2I �i = 1), we have

f

✓X

i2I

�iui

◆=X

i2I

�if (ui).

Then, for any a 2 E, the function h : E ! F givenby

h(x) = f (a + x) � f (a)

is a linear map independent of a, and

f (a + x) = f (a) + h(x), for all x 2 E.

In particular, for a = 0, if we let c = f (0), then

f (x) = c + h(x), for all x 2 E.

3.2. AFFINE MAPS 173

We should think of a as a chosen origin in E.

The function f maps the origin a in E to the origin f (a)in F .

Proposition 3.13 shows that the definition of h does notdepend on the origin chosen in E. Also, since

f (x) = c + h(x), for all x 2 E

for some fixed vector c 2 F , we see that f is the com-position of the linear map h with the translation tc (inF ).

The unique linear map h as above is called the linearmap associated with f and it is sometimes denoted by�!f .

Observe that the linear map associated with a pure trans-lation is the identity.

In view of Propositions 3.12 and 3.13, it is natural tomake the following definition.

174 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Definition 3.4. For any two vector spaces E and F , afunction f : E ! F is an a�ne map if f preserves a�necombinations, i.e., for every a�ne combination

Pi2I �iui

(withP

i2I �i = 1), we have

f

✓X

i2I

�iui

◆=X

i2I

�if (ui).

Equivalently, a function f : E ! F is an a�ne map if

there is some linear map h : E ! F (also denoted by�!f )

and some fixed vector c 2 F such that

f (x) = c + h(x), for all x 2 E.

Note that a linear map always maps the standard origin0 in E to the standard origin 0 in F .

However an a�ne map usually maps 0 to a nonzero vectorc = f (0). This is the “translation component” of thea�ne map.

3.2. AFFINE MAPS 175

When we deal with a�ne maps, it is often fruitful to thinkof the elements of E and F not only as vectors but alsoas points .

In this point of view, points can only be combined usinga�ne combinations , but vectors can be combined in anunrestricted fashion using linear combinations.

We can also think of u + v as the result of translatingthe point u by the translation tv.

These ideas lead to the definition of a�ne spaces , butthis would lead us to far afield, and for our purposes, itis enough to stick to vector spaces.

Still, one should be aware that a�ne combinations reallyapply to points, and that points are not vectors!

If E and F are finite dimensional vector spaces, withdim(E) = n and dim(F ) = m, then it is useful to repre-sent an a�ne map with respect to bases in E in F .

176 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

However, the translation part c of the a�ne map must besomewhow incorporated.

There is a standard trick to do this which amounts toviewing an a�ne map as a linear map between spaces ofdimension n + 1 and m + 1.

We also have the extra flexibility of choosing origins,a 2 E and b 2 F .

Let (u1, . . . , un) be a basis of E, (v1, . . . , vm) be a basisof F , and let a 2 E and b 2 F be any two fixed vectorsviewed as origins .

Our a�ne map f has the property that if v = f (u), then

v � b = f (a + u � a) � b = f (a) � b + h(u � a).

So, if we let y = v� b, x = u�a, and d = f (a)� b, then

y = h(x) + d, x 2 E.

3.2. AFFINE MAPS 177

Over the basis U = (u1, . . . , un), we write

x = x1u1 + · · · + xnun,

and over the basis V = (v1, . . . , vm), we write

y = y1v1 + · · · + ymvm,

d = d1v1 + · · · + dmvm.

Then, sincey = h(x) + d,

if we let A be the m ⇥ n matrix representing the linearmap h, that is, the jth column of A consists of the coor-dinates of h(uj) over the basis (v1, . . . , vm), then we canwrite

yV = AxU + dV .

where xU = (x1, . . . , xn)>, yV = (y1, . . . , ym)>, anddV = (d1, . . . , dm)>.

This is the matrix representation of our a�ne map f .

178 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

The reason for using the origins a and b is that it givesus more flexibility.

In particular, when E = F , if there is some a 2 E suchthat f (a) = a (a is a fixed point of f ), then we can pickb = a.

Then, because f (a) = a, we get

v = f (u) = f (a+u�a) = f (a)+h(u�a) = a+h(u�a),

that isv � a = h(u � a).

With respect to the new origin a, if we define x and y by

x = u � a

y = v � a,

then we gety = h(x).

Then, f really behaves like a linear map, but with respectto the new origin a (not the standard origin 0). This isthe case of a rotation around an axis that does not passthrough the origin.

3.2. AFFINE MAPS 179

Remark: A pair (a, (u1, . . . , un)) where (u1, . . . , un) isa basis of E and a is an origin chosen in E is called ana�ne frame .

We now describe the trick which allows us to incorporatethe translation part d into the matrix A.

We define the (m+1)⇥(n+1) matrix A0 obtained by firstadding d as the (n + 1)th column, and then (0, . . . , 0| {z }

n

, 1)

as the (m + 1)th row:

A0 =

✓A d0n 1

◆.

Then, it is clear that✓

y1

◆=

✓A d0n 1

◆✓x1

i↵

y = Ax + d.

180 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

This amounts to considering a point x 2 Rn as a point(x, 1) in the (a�ne) hyperplane Hn+1 in Rn+1 of equa-tion xn+1 = 1.

Then, an a�ne map is the restriction to the hyperplaneHn+1 of the linear map bf from Rn+1 to Rm+1 corre-sponding to the matrix A0, which maps Hn+1 into Hm+1

( bf (Hn+1) ✓ Hm+1).

Figure 3.2 illustrates this process for n = 2.

x1

x2

x3

(x1, x2, 1)

H3 : x3 = 1

x = (x1, x2)

Figure 3.2: Viewing Rn as a hyperplane in Rn+1 (n = 2)

3.2. AFFINE MAPS 181

For example, the map✓

x1

x2

◆7!

✓1 11 3

◆✓x1

x2

◆+

✓30

defines an a�ne map f which is represented in R3 by0

@x1

x2

1

1

A 7!

0

@1 1 31 3 00 0 1

1

A

0

@x1

x2

1

1

A .

It is easy to check that the point a = (6, �3) is fixedby f , which means that f (a) = a, so by translating thecoordinate frame to the origin a, the a�ne map behaveslike a linear map.

The idea of considering Rn as an hyperplane in Rn+1 canbe used to define projective maps .

182 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Al\ cal<) have +,ov( lejs.1h'O..{e -ro\.l( ').

I OJ¥' q cat

Figure 3.3: Dog Logic

3.3. THE DUAL SPACE E⇤ AND LINEAR FORMS 183

3.3 The Dual Space E⇤ and Linear Forms

We already observed that the field K itself (K = R orK = C) is a vector space (over itself).

The vector space Hom(E, K) of linear maps from E tothe field K, the linear forms , plays a particular role.

We take a quick look at the connection between E andE⇤ = Hom(E, K), its dual space .

As we will see shortly, every linear map f : E ! F givesrise to a linear map f> : F ⇤ ! E⇤, and it turns out thatin a suitable basis, the matrix of f> is the transpose ofthe matrix of f .

Thus, the notion of dual space provides a conceptual ex-planation of the phenomena associated with transposi-tion.

But it does more, because it allows us to view subspacesas solutions of sets of linear equations and vice-versa.

184 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Consider the following set of two “linear equations” inR3,

x � y + z = 0

x � y � z = 0,

and let us find out what is their set V of common solutions(x, y, z) 2 R3.

By subtracting the second equation from the first, we get2z = 0, and by adding the two equations, we find that2(x � y) = 0, so the set V of solutions is given by

y = x

z = 0.

This is a one dimensional subspace of R3. Geometrically,this is the line of equation y = x in the plane z = 0.

3.3. THE DUAL SPACE E⇤ AND LINEAR FORMS 185

Now, why did we say that the above equations are linear?

This is because, as functions of (x, y, z), both mapsf1 : (x, y, z) 7! x � y + z and f2 : (x, y, z) 7! x � y � zare linear.

The set of all such linear functions fromR3 toR is a vectorspace; we used this fact to form linear combinations of the“equations” f1 and f2.

Observe that the dimension of the subspace V is 1.

The ambient space has dimension n = 3 and there aretwo “independent” equations f1, f2, so it appears thatthe dimension dim(V ) of the subspace V defined by mindependent equations is

dim(V ) = n � m,

which is indeed a general fact (proved in Theorem 3.14).

186 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

More generally, in Rn, a linear equation is determined byan n-tuple (a1, . . . , an) 2 Rn, and the solutions of thislinear equation are given by the n-tuples (x1, . . . , xn) 2Rn such that

a1x1 + · · · + anxn = 0;

these solutions constitute the kernel of the linear map(x1, . . . , xn) 7! a1x1 + · · · + anxn.

The above considerations assume that we are working inthe canonical basis (e1, . . . , en) of Rn, but we can define“linear equations” independently of bases and in any di-mension, by viewing them as elements of the vector spaceHom(E, K) of linear maps from E to the field K.

3.3. THE DUAL SPACE E⇤ AND LINEAR FORMS 187

Definition 3.5. Given a vector space E, the vectorspace Hom(E, K) of linear maps from E to K is calledthe dual space (or dual) of E. The space Hom(E, K) isalso denoted by E⇤, and the linear maps in E⇤ are calledthe linear forms , or covectors . The dual space E⇤⇤ ofthe space E⇤ is called the bidual of E.

As a matter of notation, linear forms f : E ! K will alsobe denoted by starred symbol, such as u⇤, x⇤, etc.

IfE is a vector space of finite dimension n and (u1, . . . , un)is a basis of E, for any linear form f ⇤ 2 E⇤, for everyx = x1u1 + · · · + xnun 2 E, by linearity we have

f ⇤(x) = f ⇤(u1)x1 + · · · + f ⇤(nn)xn

= �1x1 + · · · + �nxn,

with �i = f ⇤(ui) 2 K for every i, 1 i n.

188 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Thus, with respect to the basis (u1, . . . , un), the linearform f ⇤ is represented by the row vector

(�1 · · · �n),

we have

f ⇤(x) =��1 · · · �n

�0

@x1...

xn

1

A ,

a linear combination of the coordinates of x, and we canview the linear form f ⇤ as a linear equation .

If we decide to use a column vector of coe�cients

c =

0

@c1...cn

1

A

instead of a row vector, then the linear form f ⇤ is definedby

f ⇤(x) = c>x.

The above notation is often used in machine learning.

3.3. THE DUAL SPACE E⇤ AND LINEAR FORMS 189

Example 3.1. Given any di↵erentiable functionf : Rn ! R, by definition, for any x 2 Rn, the totalderivative dfx of f at x is the linear form dfx : Rn ! Rdefined so that for all u = (u1, . . . , un) 2 Rn,

dfx(u) =

✓@f

@x1(x) · · · @f

@xn(x)

◆0

@u1...

un

1

A =nX

i=1

@f

@xi(x)ui.

Example 3.2. Let C([0, 1]) be the vector space of contin-uous functions f : [0, 1] ! R. The map I : C([0, 1]) ! Rgiven by

I(f ) =Z 1

0f (x)dx for any f 2 C([0, 1])

is a linear form (integration).

190 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Example 3.3. Consider the vector space Mn(R) of realn ⇥ n matrices. Let tr : Mn(R) ! R be the functiongiven by

tr(A) = a11 + a22 + · · · + ann,

called the trace of A. It is a linear form.

Let s : Mn(R) ! R be the function given by

s(A) =nX

i,j=1

aij,

where A = (aij). It is immediately verified that s is alinear form.

3.3. THE DUAL SPACE E⇤ AND LINEAR FORMS 191

Given a linear form u⇤ 2 E⇤ and a vector v 2 E, theresult u⇤(v) of applying u⇤ to v is also denoted by hu⇤, vi.

This defines a binary operation h�, �i : E⇤ ⇥ E ! Ksatisfying the following properties:

hu⇤1 + u⇤

2, vi = hu⇤1, vi + hu⇤

2, vihu⇤, v1 + v2i = hu⇤, v1i + hu⇤, v2i

h�u⇤, vi = �hu⇤, vihu⇤, �vi = �hu⇤, vi.

The above identities mean that h�, �i is a bilinear map,since it is linear in each argument.

It is often called the canonical pairing between E⇤ andE.

192 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

In view of the above identities, given any fixed vector v 2E, the map evalv : E⇤ ! K (evaluation at v) definedsuch that

evalv(u⇤) = hu⇤, vi = u⇤(v) for every u⇤ 2 E⇤

is a linear map from E⇤ to K, that is, evalv is a linearform in E⇤⇤.

Again from the above identities, the mapevalE : E ! E⇤⇤, defined such that

evalE(v) = evalv for every v 2 E,

is a linear map.

We shall see that it is injective, and that it is an isomor-phism when E has finite dimension.

3.3. THE DUAL SPACE E⇤ AND LINEAR FORMS 193

We now formalize the notion of the set V 0 of linear equa-tions vanishing on all vectors in a given subspace V ✓ E,and the notion of the set U 0 of common solutions of agiven set U ✓ E⇤ of linear equations.

The duality theorem (Theorem 3.14) shows that the di-mensions of V and V 0, and the dimensions of U and U 0,are related in a crucial way.

It also shows that, in finite dimension, the maps V 7! V 0

and U 7! U 0 are inverse bijections from subspaces of Eto subspaces of E⇤.

194 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Definition 3.6.Given a vector space E and its dual E⇤,we say that a vector v 2 E and a linear form u⇤ 2 E⇤

are orthogonal i↵ hu⇤, vi = 0. Given a subspace V ofE and a subspace U of E⇤, we say that V and U areorthogonal i↵ hu⇤, vi = 0 for every u⇤ 2 U and everyv 2 V . Given a subset V of E (resp. a subset U of E⇤),the orthogonal V 0 of V is the subspace V 0 of E⇤ definedsuch that

V 0 = {u⇤ 2 E⇤ | hu⇤, vi = 0, for every v 2 V }

(resp. the orthogonal U 0 of U is the subspace U 0 of Edefined such that

U 0 = {v 2 E | hu⇤, vi = 0, for every u⇤ 2 U}).

The subspace V 0 ✓ E⇤ is also called the annihilator ofV .

3.3. THE DUAL SPACE E⇤ AND LINEAR FORMS 195

The subspace U 0 ✓ E annihilated by U ✓ E⇤ does nothave a special name. It seems reasonable to call it thelinear subspace (or linear variety) defined by U .

Informally, V 0 is the set of linear equations that vanishon V , and U 0 is the set of common zeros of all linearequations in U . We can also define V 0 by

V 0 = {u⇤ 2 E⇤ | V ✓ Keru⇤}

and U 0 by

U 0 =\

u⇤2U

Keru⇤.

Observe that E0 = {0} = (0), and {0}0 = E⇤.

Furthermore, if V1 ✓ V2 ✓ E, then V 02 ✓ V 0

1 ✓ E⇤, andif U1 ✓ U2 ✓ E⇤, then U 0

2 ✓ U 01 ✓ E.

196 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

It can also be shown that that V ✓ V 00 for every sub-space V of E, and that U ✓ U 00 for every subspace U ofE⇤.

We will see shortly that in finite dimension, we have

V = V 00 and U = U 00.

Here are some examples. Let E = M2(R), the space ofreal 2 ⇥ 2 matrices, and let V be the subspace of M2(R)spanned by the matrices

✓0 11 0

◆,

✓1 00 0

◆,

✓0 00 1

◆.

We check immediately that the subspace V consists of allmatrices of the form

✓b aa c

◆,

that is, all symmetric matrices.

3.3. THE DUAL SPACE E⇤ AND LINEAR FORMS 197

The matrices✓

a11 a12

a21 a22

in V satisfy the equation

a12 � a21 = 0,

and all scalar multiples of these equations, so V 0 is thesubspace of E⇤ spanned by the linear form given by

u⇤(a11, a12, a21, a22) = a12 � a21.

By the duality theorem (Theorem 3.14) we have

dim(V 0) = dim(E) � dim(V ) = 4 � 3 = 1.

The above example generalizes to E = Mn(R) for anyn � 1, but this time, consider the space U of linear formsasserting that a matrix A is symmetric; these are thelinear forms spanned by the n(n � 1)/2 equations

aij � aji = 0, 1 i < j n;

198 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Note there are no constraints on diagonal entries, and halfof the equations

aij � aji = 0, 1 i 6= j n

are redundant. It is easy to check that the equations(linear forms) for which i < j are linearly independent.

To be more precise, let U be the space of linear forms inE⇤ spanned by the linear forms

u⇤ij(a11, . . . , a1n, a21, . . . , a2n, . . . , an1, . . . , ann)

= aij � aji, 1 i < j n.

The dimension of U is n(n � 1)/2. Then, the set U 0 ofcommon solutions of these equations is the space S(n) ofsymmetric matrices.

By the duality theorem (Theorem 3.14), this space hasdimension

n(n + 1)

2= n2 � n(n � 1)

2.

3.3. THE DUAL SPACE E⇤ AND LINEAR FORMS 199

If E = Mn(R), consider the subspace U of linear formsin E⇤ spanned by the linear forms

u⇤ij(a11, . . . , a1n, a21, . . . , a2n, . . . , an1, . . . , ann)

= aij + aji, 1 i < j n

u⇤ii(a11, . . . , a1n, a21, . . . , a2n, . . . , an1, . . . , ann)

= aii, 1 i n.

It is easy to see that these linear forms are linearly inde-pendent, so dim(U) = n(n + 1)/2.

The space U 0 of matrices A 2 Mn(R) satifying all of theabove equations is clearly the space Skew(n) of skew-symmetric matrices.

By the duality theorem (Theorem 3.14), the dimension ofU 0 is

n(n � 1)

2= n2 � n(n + 1)

2.

200 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

For yet another example with E = Mn(R), for any A 2Mn(R), consider the linear form in E⇤ given by

tr(A) = a11 + a22 + · · · + ann,

called the trace of A.

The subspace U 0 of E consisting of all matrices A suchthat tr(A) = 0 is a space of dimension n2 � 1.

The dimension equations

dim(V ) + dim(V 0) = dim(E)

dim(U) + dim(U 0) = dim(E)

are always true (if E is finite-dimensional). This is partof the duality theorem (Theorem 3.14).

In constrast with the previous examples, given a matrixA 2 Mn(R), the equations asserting that A>A = I arenot linear constraints.

3.3. THE DUAL SPACE E⇤ AND LINEAR FORMS 201

For example, for n = 2, we have

a211 + a2

21 = 1

a221 + a2

22 = 1

a11a12 + a21a22 = 0.

Given a vector space E and any basis (ui)i2I for E, wecan associate to each ui a linear form u⇤

i 2 E⇤, and theu⇤

i have some remarkable properties.

Definition 3.7. Given a vector space E and any basis(ui)i2I for E, by Proposition 1.10, for every i 2 I , thereis a unique linear form u⇤

i such that

u⇤i (uj) =

⇢1 if i = j0 if i 6= j,

for every j 2 I . The linear form u⇤i is called the coordi-

nate form of index i w.r.t. the basis (ui)i2I .

202 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Remark: Given an index set I , authors often define theso called Kronecker symbol �i j, such that

�i j =

⇢1 if i = j0 if i 6= j,

for all i, j 2 I .

Then,u⇤

i (uj) = �i j.

The reason for the terminology coordinate form is asfollows: If E has finite dimension and if (u1, . . . , un) is abasis of E, for any vector

v = �1u1 + · · · + �nun,

we have

u⇤i (v) = �i.

Therefore, u⇤i is the linear function that returns the ith co-

ordinate of a vector expressed over the basis (u1, . . . , un).

As part of the Duality theorem (Theorem 3.14), we willprove that if (u1, . . . , un) is a basis of E, then (u⇤

1, . . . , u⇤n)

is a basis of E⇤ called the dual basis .

3.3. THE DUAL SPACE E⇤ AND LINEAR FORMS 203

If (u1, . . . , un) is a basis of Rn (more generally Kn), itis possible to find explicitly the dual basis (u⇤

1, . . . , u⇤n),

where each u⇤i is represented by a row vector.

For example, consider the columns of the Bezier matrix

B4 =

0

BB@

1 �3 3 �10 3 �6 30 0 3 �30 0 0 1

1

CCA .

The form u⇤1 is represented by a row vector (�1 �2 �3 �4)

such that

��1 �2 �3 �4

0

BB@

1 �3 3 �10 3 �6 30 0 3 �30 0 0 1

1

CCA =�1 0 0 0

�.

This implies that u⇤1 is the first row of the inverse of B4.

204 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Since

B�14 =

0

BB@

1 1 1 10 1/3 2/3 10 0 1/3 10 0 0 1

1

CCA ,

the linear forms (u⇤1, u

⇤2, u

⇤3, u

⇤4) correspond to the rows of

B�14 .

In particular, u⇤1 is represented by (1 1 1 1).

The above method works for any n. Given any basis(u1, . . . , un) of Rn, if P is the n ⇥ n matrix whose jthcolumn is uj, then the dual form u⇤

i is given by the ithrow of the matrix P�1.

We have the following important duality theorem adaptedfrom E. Artin.

3.3. THE DUAL SPACE E⇤ AND LINEAR FORMS 205

Theorem 3.14. (Duality theorem) Let E be a vectorspace of dimension n. The following properties hold:

(a) For every basis (u1, . . . , un) of E, the family of co-ordinate forms (u⇤

1, . . . , u⇤n) is a basis of E⇤.

(b) For every subspace V of E, we have V 00 = V .

(c) For every pair of subspaces V and W of E suchthat E = V �W , with V of dimension m, for everybasis (u1, . . . , un) of E such that (u1, . . . , um) is abasis of V and (um+1, . . . , un) is a basis of W , thefamily (u⇤

1, . . . , u⇤m) is a basis of the orthogonal W 0

of W in E⇤. Furthermore, we have W 00 = W , and

dim(W ) + dim(W 0) = dim(E).

(d) For every subspace U of E⇤, we have

dim(U) + dim(U 0) = dim(E),

where U 0 is the orthogonal of U in E, andU 00 = U .

206 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Part (a) of Theorem 3.14 shows that

dim(E) = dim(E⇤),

and if (u1, . . . , un) is a basis of E, then (u⇤1, . . . , u

⇤n) is

a basis of the dual space E⇤ called the dual basis of(u1, . . . , un).

Define the function E from subspaces of E to subspaces ofE⇤ and the function Z from subspaces of E⇤ to subspacesof E by

E(V ) = V 0, V ✓ E

Z(U) = U 0, U ✓ E⇤.

By part (c) and (d) of theorem 3.14,

(Z � E)(V ) = V 00 = V

(E � Z)(U) = U 00 = U,

so Z �E = id and E �Z = id, and the maps E and V areinverse bijections.

3.3. THE DUAL SPACE E⇤ AND LINEAR FORMS 207

These maps set up a duality between subspaces of E, andsubspaces of E⇤.

� One should be careful that this bijection does not holdif E has infinite dimension. Some restrictions on the

dimensions of U and V are needed.

Suppose that V is a subspace of Rn of dimension m andthat (v1, . . . , vm) is a basis of V .

To find a basis of V 0, we first extend (v1, . . . , vm) to abasis (v1, . . . , vn) of Rn, and then by part (c) of Theorem3.14, we know that (v⇤

m+1, . . . , v⇤n) is a basis of V 0.

For example, suppose that V is the subspace ofR4 spannedby the two linearly independent vectors

v1 =

0

BB@

1111

1

CCA v2 =

0

BB@

11

�1�1

1

CCA ,

the first two vectors of the Haar basis in R4.

208 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

The four columns of the Haar matrix

W =

0

BB@

1 1 1 01 1 �1 01 �1 0 11 �1 0 �1

1

CCA

form a basis of R4, and the inverse of W is given by

W�1 =

0

BB@

1/4 1/4 1/4 1/41/4 1/4 �1/4 �1/41/2 �1/2 0 00 0 1/2 �1/2

1

CCA .

Since the dual basis (v⇤1, v

⇤2, v

⇤3, v

⇤4) is given by the row of

W�1, the last two rows of W�1,✓1/2 �1/2 0 00 0 1/2 �1/2

◆,

form a basis of V 0.

3.3. THE DUAL SPACE E⇤ AND LINEAR FORMS 209

We also obtain a basis by rescaling by the factor 1/2, sothe linear forms given by the row vectors

✓1 �1 0 00 0 1 �1

form a basis of V 0, the space of linear forms (linear equa-tions) that vanish on the subspace V .

The method that we described to find V 0 requires firstextending a basis of V and then inverting a matrix, butthere is a more direct method.

210 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Indeed, let A be the n ⇥ m matrix whose columns arethe basis vectors (v1, . . . , vm) of V . Then, a linear formu represented by a row vector belongs to V 0 i↵ uvi = 0for i = 1, . . . , m i↵

uA = 0

i↵

A>u> = 0.

Therefore, all we need to do is to find a basis of thenullspace of A>.

This can be done quite e↵ectively using the reduction ofa matrix to reduced row echelon form (rref); see Section4.5.

3.3. THE DUAL SPACE E⇤ AND LINEAR FORMS 211

Here is another example illustrating the power of Theo-rem 3.14.

Let E = Mn(R), and consider the equations assertingthat the sum of the entries in every row of a matrix 2Mn(R) is equal to the same number.

We have n � 1 equations

nX

j=1

(aij � ai+1j) = 0, 1 i n � 1,

and it is easy to see that they are linearly independent.

Therefore, the space U of linear forms in E⇤ spanned bythe above linear forms (equations) has dimension n � 1,and the space U 0 of matrices sastisfying all these equa-tions has dimension n2 � n + 1.

It is not so obvious to find a basis for this space.

212 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

When E is of finite dimension n and (u1, . . . , un) is abasis of E, we noted that the family (u⇤

1, . . . , u⇤n) is a

basis of the dual space E⇤,

Let us see how the coordinates of a linear form '⇤ 2 E⇤

over the basis (u⇤1, . . . , u

⇤n) vary under a change of basis.

Let (u1, . . . , un) and (v1, . . . , vn) be two bases of E, andlet P = (ai j) be the change of basis matrix from (u1, . . . , un)to (v1, . . . , vn), so that

vj =nX

i=1

ai jui.

If

'⇤ =nX

i=1

'iu⇤i =

nX

i=1

'0iv

⇤i ,

after some algebra, we get

'0j =

nX

i=1

ai j'i.

3.3. THE DUAL SPACE E⇤ AND LINEAR FORMS 213

Comparing with the change of basis

vj =nX

i=1

ai jui,

we note that this time, the coordinates ('i) of the linearform '⇤ change in the same direction as the change ofbasis.

For this reason, we say that the coordinates of linear formsare covariant .

By abuse of language, it is often said that linear formsare covariant , which explains why the term covector isalso used for a linear form.

Observe that if (e1, . . . , en) is a basis of the vector spaceE, then, as a linear map from E to K, every linear formf 2 E⇤ is represented by a 1 ⇥ n matrix, that is, by arow vector

(�1 · · · �n),

with respect to the basis (e1, . . . , en) of E, and 1 of K,where f (ei) = �i.

214 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

A vector u =Pn

i=1 uiei 2 E is represented by a n ⇥ 1matrix, that is, by a column vector

0

@u1...

un

1

A ,

and the action of f on u, namely f (u), is represented bythe matrix product

��1 · · · �n

�0

@u1...

un

1

A = �1u1 + · · · + �nun.

On the other hand, with respect to the dual basis (e⇤1, . . . , e

⇤n)

of E⇤, the linear form f is represented by the column vec-tor

0

@�1...

�n

1

A .

3.3. THE DUAL SPACE E⇤ AND LINEAR FORMS 215

We will now pin down the relationship between a vectorspace E and its bidual E⇤⇤.

Proposition 3.15. Let E be a vector space. The fol-lowing properties hold:

(a) The linear map evalE : E ! E⇤⇤ defined such that

evalE(v) = evalv, for all v 2 E,

that is, evalE(v)(u⇤) = hu⇤, vi = u⇤(v) for everyu⇤ 2 E⇤, is injective.

(b) When E is of finite dimension n, the linear mapevalE : E ! E⇤⇤ is an isomorphism (called thecanonical isomorphism).

When E is of finite dimension and (u1, . . . , un) is a basisof E, in view of the canonical isomorphismevalE : E ! E⇤⇤, the basis (u⇤⇤

1 , . . . , u⇤⇤n ) of the bidual is

identified with (u1, . . . , un).

Proposition 3.15 can be reformulated very fruitfully interms of pairings.

216 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Definition 3.8. Given two vector spaces E and F overK, a pairing between E and F is a bilinear map' : E ⇥ F ! K. Such a pairing is nondegenerate i↵

(1) for every u 2 E, if '(u, v) = 0 for all v 2 F , thenu = 0, and

(2) for every v 2 F , if '(u, v) = 0 for all u 2 E, thenv = 0.

A pairing ' : E ⇥ F ! K is often denoted byh�, �i : E ⇥ F ! K.

For example, the map h�, �i : E⇤ ⇥ E ! K definedearlier is a nondegenerate pairing (use the proof of (a) inProposition 3.15).

Given a pairing ' : E ⇥F ! K, we can define two mapsl' : E ! F ⇤ and r' : F ! E⇤ as follows:

3.3. THE DUAL SPACE E⇤ AND LINEAR FORMS 217

For every u 2 E, we define the linear form l'(u) in F ⇤

such that

l'(u)(y) = '(u, y) for every y 2 F ,

and for every v 2 F , we define the linear form r'(v) inE⇤ such that

r'(v)(x) = '(x, v) for every x 2 E.

Proposition 3.16.Given two vector spaces E and Fover K, for every nondegenerate pairing' : E ⇥ F ! K between E and F , the mapsl' : E ! F ⇤ and r' : F ! E⇤ are linear and injec-tive. Furthermore, if E and F have finite dimension,then this dimension is the same and l' : E ! F ⇤ andr' : F ! E⇤ are bijections.

When E has finite dimension, the nondegenerate pair-ing h�, �i : E⇤ ⇥ E ! K yields another proof of theexistence of a natural isomorphism between E and E⇤⇤.

Interesting nondegenerate pairings arise in exterior alge-bra.

218 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

ME,-rt< Ie. C l-OCK

Figure 3.4: Metric Clock

3.4. HYPERPLANES AND LINEAR FORMS 219

3.4 Hyperplanes and Linear Forms

Actually, Proposition 3.17 below follows from parts (c)and (d) of Theorem 3.14, but we feel that it is also inter-esting to give a more direct proof.

Proposition 3.17. Let E be a vector space. The fol-lowing properties hold:

(a) Given any nonnull linear form f ⇤ 2 E⇤, its kernelH = Ker f ⇤ is a hyperplane.

(b) For any hyperplane H in E, there is a (nonnull)linear form f ⇤ 2 E⇤ such that H = Ker f ⇤.

(c) Given any hyperplane H in E and any (nonnull)linear form f ⇤ 2 E⇤ such that H = Ker f ⇤, forevery linear form g⇤ 2 E⇤, H = Ker g⇤ i↵ g⇤ = �f ⇤

for some � 6= 0 in K.

We leave as an exercise the fact that every subspaceV 6= E of a vector space E, is the intersection of allhyperplanes that contain V .

We now consider the notion of transpose of a linear mapand of a matrix.

220 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

3.5 Transpose of a Linear Map and of a Matrix

Given a linear map f : E ! F , it is possible to define amap f> : F ⇤ ! E⇤ which has some interesting proper-ties.

Definition 3.9. Given a linear map f : E ! F , thetranspose f> : F ⇤ ! E⇤ of f is the linear map definedsuch that

f>(v⇤) = v⇤ � f,

for every v⇤ 2 F ⇤.

Equivalently, the linear map f> : F ⇤ ! E⇤ is definedsuch that

hv⇤, f (u)i = hf>(v⇤), ui,

for all u 2 E and all v⇤ 2 F ⇤.

3.5. TRANSPOSE OF A LINEAR MAP AND OF A MATRIX 221

It is easy to verify that the following properties hold:

(f + g)> = f> + g>

(g � f )> = f> � g>

id>E = idE⇤.

� Note the reversal of composition on the right-hand sideof (g � f )> = f> � g>.

The equation (g � f )> = f> � g> implies the followinguseful proposition.

Proposition 3.18. If f : E ! F is any linear map,then the following properties hold:

(1) If f is injective, then f> is surjective.

(2) If f is surjective, then f> is injective.

222 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

We also have the following property showing the natural-ity of the eval map.

Proposition 3.19. For any linear map f : E ! F ,we have

f>> � evalE = evalF � f,

or equivalently, the following diagram commutes:

E⇤⇤ f>>//F ⇤⇤

E

evalE

OO

f//F.

evalF

OO

3.5. TRANSPOSE OF A LINEAR MAP AND OF A MATRIX 223

If E and F are finite-dimensional, then evalE and evalFare isomorphisms, so Proposition 3.19 shows that

f>> = eval�1F � f � evalE. (⇤)

The above equation is often interpreted as follows: if weidentify E with its bidual E⇤⇤ and F with its bidual F ⇤⇤,then f>> = f .

This is an abuse of notation; the rigorous statement is(⇤).

Proposition 3.20. Given a linear map f : E ! F ,for any subspace V of E, we have

f (V )0 = (f>)�1(V 0) = {w⇤ 2 F ⇤ | f>(w⇤) 2 V 0}.

As a consequence,

Ker f> = (Im f )0 and Ker f = (Im f>)0.

224 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

The following theorem shows the relationship between therank of f and the rank of f>.

Theorem 3.21. Given a linear map f : E ! F , thefollowing properties hold.

(a) The dual (Im f )⇤ of Im f is isomorphic toIm f> = f>(F ⇤); that is,

(Im f )⇤ ⇡ Im f>.

(b) If F is finite dimensional, then rk(f ) = rk(f>).

The following proposition can be shown, but it requires ageneralization of the duality theorem.

Proposition 3.22. If f : E ! F is any linear map,then the following identities hold:

Im f> = (Ker (f ))0

Ker (f>) = (Im f )0

Im f = (Ker (f>)0

Ker (f ) = (Im f>)0.

3.5. TRANSPOSE OF A LINEAR MAP AND OF A MATRIX 225

The following proposition shows the relationship betweenthe matrix representing a linear map f : E ! F and thematrix representing its transpose f> : F ⇤ ! E⇤.

Proposition 3.23. Let E and F be two vector spaces,and let (u1, . . . , un) be a basis for E, and (v1, . . . , vm)be a basis for F . Given any linear map f : E ! F ,if M(f ) is the m ⇥ n-matrix representing f w.r.t.the bases (u1, . . . , un) and (v1, . . . , vm), the n ⇥ m-matrix M(f>) representing f> : F ⇤ ! E⇤ w.r.t. thedual bases (v⇤

1, . . . , v⇤m) and (u⇤

1, . . . , u⇤n) is the trans-

pose M(f )> of M(f ).

We now can give a very short proof of the fact that therank of a matrix is equal to the rank of its transpose.

Proposition 3.24. Given a m ⇥ n matrix A over afield K, we have rk(A) = rk(A>).

Thus, given an m ⇥ n-matrix A, the maximum numberof linearly independent columns is equal to the maximumnumber of linearly independent rows.

226 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Proposition 3.25.Given any m⇥n matrix A over afield K (typically K = R or K = C), the rank of A isthe maximum natural number r such that there is aninvertible r ⇥ r submatrix of A obtained by selectingr rows and r columns of A.

For example, the 3 ⇥ 2 matrix

A =

0

@a11 a12

a21 a22

a31 a32

1

A

has rank 2 i↵ one of the three 2 ⇥ 2 matrices✓

a11 a12

a21 a22

◆ ✓a11 a12

a31 a32

◆ ✓a21 a22

a31 a32

is invertible. We will see in Chapter 5 that this is equiv-alent to the fact the determinant of one of the abovematrices is nonzero.

This is not a very e�cient way of finding the rank ofa matrix. We will see that there are better ways usingvarious decompositions such as LU, QR, or SVD.

3.5. TRANSPOSE OF A LINEAR MAP AND OF A MATRIX 227

OF IH1S.1 S IHkf rf '5 OAJL'f OF IMPoRTArJc.E:., AND 1FfE:RE. IS NO WAy

rf cAN SS or MY pgAC:lltAL use. \\

Figure 3.5: Beauty

228 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

3.6 The Four Fundamental Subspaces

Given a linear map f : E ! F (where E and F arefinite-dimensional), Proposition 3.20 revealed that thefour spaces

Im f, Im f>, Ker f, Ker f>

play a special role. They are often called the fundamentalsubspaces associated with f .

These spaces are related in an intimate manner, sinceProposition 3.20 shows that

Ker f = (Im f>)0

Ker f> = (Im f )0,

and Theorem 3.21 shows that

rk(f ) = rk(f>).

3.6. THE FOUR FUNDAMENTAL SUBSPACES 229

It is instructive to translate these relations in terms ofmatrices (actually, certain linear algebra books make abig deal about this!).

If dim(E) = n and dim(F ) = m, given any basis (u1, . . .,un) of E and a basis (v1, . . . , vm) of F , we know that f isrepresented by an m⇥n matrix A = (ai j), where the jthcolumn of A is equal to f (uj) over the basis (v1, . . . , vm).

Furthermore, the transpose map f> is represented by then ⇥ m matrix A> (with respect to the dual bases).

Consequently, the four fundamental spaces

Im f, Im f>, Ker f, Ker f>

correspond to

230 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

(1) The column space of A, denoted by ImA or R(A);this is the subspace of Rm spanned by the columns ofA, which corresponds to the image Im f of f .

(2) The kernel or nullspace of A, denoted by KerA orN (A); this is the subspace of Rn consisting of allvectors x 2 Rn such that Ax = 0.

(3) The row space of A, denoted by ImA> or R(A>);this is the subspace of Rn spanned by the rows of A,or equivalently, spanned by the columns of A>, whichcorresponds to the image Im f> of f>.

(4) The left kernel or left nullspace of A denoted byKerA> orN (A>); this is the kernel (nullspace) ofA>,the subspace of Rm consisting of all vectors y 2 Rm

such that A>y = 0, or equivalently, y>A = 0.

Recall that the dimension r of Im f , which is also equalto the dimension of the column space ImA = R(A), isthe rank of A (and f ).

3.6. THE FOUR FUNDAMENTAL SUBSPACES 231

Then, some our previous results can be reformulated asfollows:

1. The column space R(A) of A has dimension r.

2. The nullspace N (A) of A has dimension n � r.

3. The row space R(A>) has dimension r.

4. The left nullspace N (A>) of A has dimension m � r.

The above statements constitute what Strang calls theFundamental Theorem of Linear Algebra, Part I (seeStrang [32]).

232 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

The two statements

Ker f = (Im f>)0

Ker f> = (Im f )0

translate to

(1) The nullspace of A is the orthogonal of the row spaceof A.

(2) The left nullspace of A is the orthogonal of the columnspace of A.

The above statements constitute what Strang calls theFundamental Theorem of Linear Algebra, Part II (seeStrang [32]).

Since vectors are represented by column vectors and linearforms by row vectors (over a basis in E or F ), a vectorx 2 Rn is orthogonal to a linear form y if

yx = 0.

3.6. THE FOUR FUNDAMENTAL SUBSPACES 233

Then, a vector x 2 Rn is orthogonal to the row space ofA i↵ x is orthogonal to every row of A, namelyAx = 0, which is equivalent to the fact that x belong tothe nullspace of A.

Similarly, the column vector y 2 Rm (representing alinear form over the dual basis of F ⇤) belongs to thenullspace of A> i↵ A>y = 0, i↵ y>A = 0, which meansthat the linear form given by y> (over the basis in F ) isorthogonal to the column space of A.

Since (2) is equivalent to the fact that the column spaceof A is equal to the orthogonal of the left nullspace ofA, we get the following criterion for the solvability of anequation of the form Ax = b:

The equation Ax = b has a solution i↵ for all y 2 Rm, ifA>y = 0, then y>b = 0.

234 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

Indeed, the condition on the right-hand side says that bis orthogonal to the left nullspace of A, that is, that bbelongs to the column space of A.

This criterion can be cheaper to check that checking di-rectly that b is spanned by the columns of A. For exam-ple, if we consider the system

x1 � x2 = b1

x2 � x3 = b2

x3 � x1 = b3

which, in matrix form, is written Ax = b as below:0

@1 �1 00 1 �1

�1 0 1

1

A

0

@x1

x2

x3

1

A =

0

@b1

b2

b3

1

A ,

we see that the rows of the matrix A add up to 0.

3.6. THE FOUR FUNDAMENTAL SUBSPACES 235

In fact, it is easy to convince ourselves that the left nullspaceof A is spanned by y = (1, 1, 1), and so the system is solv-able i↵ y>b = 0, namely

b1 + b2 + b3 = 0.

Note that the above criterion can also be stated negativelyas follows:

The equation Ax = b has no solution i↵ there is somey 2 Rm such that A>y = 0 and y>b 6= 0.

236 CHAPTER 3. DIRECT SUMS, AFFINE MAPS, THE DUAL SPACE, DUALITY

--- ---

.!(</I' _</II - 'leJ'A dX)at 1£ I

I 2eV = - i (p:, - PI) • h'

Figure 3.6: Brain Size?