0% found this document useful (0 votes)
10 views20 pages

Chapter_4

Uploaded by

ramzyputra05
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
Download as pdf or txt
0% found this document useful (0 votes)
10 views20 pages

Chapter_4

Uploaded by

ramzyputra05
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
Download as pdf or txt
Download as pdf or txt
You are on page 1/ 20

Chapter 4

Linear maps

Before concentrating on linear maps, we provide a more general setting.

4.1 General maps


We start with the general definition of a map between two sets, and introduce some
notations.
Definition 4.1.1. Let S, S ′ be two sets. A map T from S to S ′ is a rule which associates
to each element of S an element of S ′ . The notation

T : S ∋ X 7→ T(X) ∈ S ′

will be used for such a map. If X ∈ S, then T(X) ∈ S ′ is called the image of X by T.
The set S is often called the domain of T and is also denoted by Dom(T), while
{ }
T(S) := T(X) | X ∈ S

is often called the range of T and is also denoted by Ran(T).


Examples 4.1.2. (i) The function f : R ∋ x 7→ f (x) = x2 − 3x + 2 ∈ R is a map
from R to R,

(ii) Any A ∈ Mmn (R) defines a map LA : Rn → Rm by LA (X) := AX for any


X ∈ Rn . More generally, for any field F and any A ∈ Mmn (F), one defines a map
LA : Fn → Fm by LA (X) := AX for any X ∈ Fn ,
(x) (x) ( 2 )
(iii) The rule F : R3 ∋ yz 7→ F yz = x+y+z+3 x +y
∈ R2 is a map,

(iv) Let C 1 (R) := {continuous functions f on R | f ′ exists and is continuous } and


let C(R) := {continuous functions f on R}. Then the following rule defines a
map:
D : C 1 (R) ∋ f 7→ Df = f ′ ∈ C(R)

61
62 CHAPTER 4. LINEAR MAPS

(v) For any A, B ∈ Mn (R) one can define a map by

TA,B : Mn (R) ∋ X 7→ TA,B (X) = AX + B ∈ Mn (R),

(vi) The function g : R∗ ∋ x 7→ g(x) = 3x−2


x
∈ R is a map from R∗ to R, but is not a
map from R to R because g(0) is not defined,

(vii) For any fixed Y ∈ Rn , a map is defined by TY : Rn ∋ X 7→ TY (X) = X +Y ∈ Rn ,


and is called the translation by Y .

Remark 4.1.3. For a map T : S → S ′ , the determination of Ran(T) is not always an


easy task. For example if one considers f : R → R with f (x) = x2 − 3x + 2, then one
has to look for the minimum of f , which is −1/4 obtained for x = 3/2, and one can
then set Ran(f ) = [−1/4, ∞). Similarly, if A ∈ Mmn (R), then what is the range of LA ,
i.e. the set of Y ∈ Rm such that Y = AX for some X ∈ Rn ?

We end this section with a natural definition.

Definition 4.1.4. Let T : S → S ′ be a map, let W ⊂ S be a subset of S and let Z be


a subset of S ′ . Then the set T(W ) := {T(X) | X ∈ W } is called the image of W by T,
while the set { }
T−1 (Z) := X ∈ S | T(X) ∈ Z
is called the preimage of Z by T.

4.2 Linear maps


From now on, we shall concentrate on the simplest maps, the linear ones. Note that in
order to state the next definition, one has to deal with vector spaces instead of arbitrary
sets, and in addition the two vector spaces have to be defined on the same field.

Definition 4.2.1. Let V, W be two vector spaces over the same field F. A map T : V →
W is a linear map if the following two conditions are satisfied:

(i) T(X + Y ) = T(X) + T(Y ) for any X, Y ∈ V ,

(ii) T(λX) = λT(X) for any X ∈ V and λ ∈ F.

Note that the examples (ii) and (iv) of Examples 4.1.2 were already linear maps.
Let us still mention the map Id : V → V (also denoted by 1) defined by Id(X) = X for
any X ∈ V , which is clearly linear, and the map O : V → W defined by O(X) = 0 for
any X ∈ V , which is also linear.
Let us now observe that linear maps are rather simple maps.

Lemma 4.2.2. Let V, W be vector spaces over the same field F, and let T : V → W be
a linear map. Then,
4.2. LINEAR MAPS 63

(i) T(0) = 0,
(ii) T(−X) = −T(X) for any X ∈ V .
Proof. (i) It is sufficient to observe that

T(0) = T(0 + 0) = T(0) + T(0) = 2T(0)

which implies the result.


(i) Observe that

0 = T(0) = T(X − X) = T(X) + T(−X)

which directly leads to the result.


Let go us step further in abstraction and consider families of linear maps. For that
purpose, let us first define an addition of linear maps, and the multiplication of a linear
map by a scalar. Namely, if V, W are vector spaces over the same field F and if T1 , T2
are linear maps from V to W , one sets
( )
T1 + T2 (X) = T1 (X) + T2 (X) for any X ∈ V. (4.2.1)

If λ ∈ F and if T : V → W is linear, one also sets


( )
λT (X) = λT(X) for any X ∈ V. (4.2.2)

It is then easily observed that T1 + T2 is still a linear map, and that λT is also a linear
map. We can then even say more:
Proposition 4.2.3. Let V, W be vector spaces over the same field F. Then
{ }
L(V, W ) := T : V → W | T is linear ,

is a vector space over F, once endowed with the addition defined by (4.2.1) and the
multiplication by a scalar defined in (4.2.2).
Before giving the proof, let us observe that if V = Rn and W = Rn , then L(Rn , Rm )
corresponds to the set of all LA with A ∈ Mmn (R). Note that this statement also holds
for arbitrary field F, i.e.
{ }
L(Fn , Fm ) = LA | A ∈ Mmn (F) .

Proof. The proof consists in checking all conditions of Definition 3.1.3. For that purpose,
consider T, T1 , T2 , T3 be linear maps from V to W , and let λ, µ ∈ F. Let also X be an
arbitrary element of V .
(i) One has
[ ]
(T1 + T2 ) + T3 (X) = (T1 + T2 )(X) + T3 (X) = T1 (X) + T2 (X) + T3 (X)
[ ]
= T1 (X) + (T2 + T3 )(X) = T1 + (T2 + T3 ) (X).
64 CHAPTER 4. LINEAR MAPS

Since X is arbitrary, it follows that (T1 + T2 ) + T3 = T1 + (T2 + T3 ).


(ii) One has
(T1 + T2 )(X) = T1 (X) + T2 (X) = T2 (X) + T1 (X) = (T2 + T1 )(X)
which implies that T1 + T2 = T2 + T1 .
(iii) We already know that O : V → W is linear, which means that O ∈ L(V, W ).
In addition, one clearly has T + O = O + T = T.
(iv) By setting [−T](X) := −T(X), one readily observes that −T ∈ L(V, W ) and
by using the addition (4.2.1) one infers that T + (−T) = O.
(v) Similarly, λT ∈ L(V, W ) and 1T = T.
The remaining three properties are easily checked by using the definition 4.2.2 and
the basic properties of vector spaces.

Question : If dim(V ) = n and if dim(W ) = m, what is the dimension of L(V, W ) ?


Let us now consider a linear map T : V → Rn with V a real vector space. Since for
each X ∈ V one has T(X) ∈ Rn , one often sets
 
T1 (X)
 
T(X) =  ...  (4.2.3)
Tn (X)
with Tj (X) := T(X)j the j th component of T evaluated at X. Thus, T defines a family
of maps Tj : V → R, and reciprocally, any family {Tj }nj=1 with Tj : V → R defines a
map T : V → Rn by (4.2.3). Sometimes, the maps T1 , . . . , Tn are called the components
of T.
Example 4.2.4. If T : R2 → R3 is defined by
 
( ) 2x − y
x
T = 3x + 4y  ,
y
x − 5y
( ) ( ) ( )
then T = t (T1 , T2 , T3 ) with T1 xy = 2x − y, T2 xy = 3x + 4y and T3 xy = x − 5y.
More generally:
Lemma 4.2.5. Let V be a vector space over a field F, and let T : V → Fn with
T = t (T1 , . . . , Tn ) the components of T. Then T is a linear map if and only if each Tj
is a linear map.
( )
Proof.
( One has T(X + Y ) = t T1 (X) + Y ), . . . , Tn (X + Y ) and T(X) + T(Y ) =
t
T1 (X) + T1 (Y ), . . . , Tn (X) + Tn (Y ) . It then follows that
   
T1 (X + Y ) T1 (X) + T1 (Y )
 ..   .. 
T(X + Y ) = T(X) + T(Y ) ⇐⇒  . = . ,
Tn (X + Y ) Tn (X) + Tn (Y )
which corresponds to half of the statement. A similar argument holds for the multipli-
cation by a scalar.
4.3. KERNEL AND RANGE OF A LINEAR MAP 65

4.3 Kernel and range of a linear map


Let V, W be two vector spaces over the same field F and let T : V → W be a linear
map. Recall that

Ran(T) := {Y ∈ W | Y = T(X) for some X ∈ V }

and
Ker(T) := {X ∈ V | T(X) = 0}.

Lemma 4.3.1. In the previous setting, Ker(T) is a subspace of V while Ran(T) is a


subspace of W .

Proof. The first part of the statement is proved in Exercise 4.4. For the second part
of the statement, consider Y1 , Y2 ∈ Ran(T), i.e. there exist X1 , X2 ∈ V such that
Y1 = T(X1 ) and Y2 = T(X2 ). Then one has

Y1 + Y2 = T(X1 ) + T(X2 ) = T(X1 + X2 )

with X1 + X2 ∈ V . In other words, Y1 + Y2 belongs to Ran(T). Similarly, for λ ∈ F and


any Y = T(X) with X ∈ V one has

λY = λT(X) = T(λX)

with λX ∈ V . Again, it follows that λY ∈ Ran(T), from which one concludes that
Ran(T) is a subspace of W .

Examples 4.3.2. (i) Let N ∈ Rn with N ̸= 0, and let us set TN : Rn → R by


TN (X) = N · X. In this case, TN is a linear map. Indeed, one has

TN (X + Y ) = N · (X + Y ) = N · X + N · Y = TN (X) + TN (Y ),

and similarly TN (λX) = N · (λX) = λ(N · X) = λTN (X). Then one observes
that

Ker(TN ) = {X ∈ Rn | N · X = 0} = {X ∈ Rn | X · N = 0 · N } = HN,0 .

On the other hand, Ran(TN ) = R, as it can easily be checked by considering


elements X of the form λN , for any λ ∈ R.

(ii) Let A ∈ Mmn (R) and let us set LA : Rn → Rm defined by LA (X) = AX for any
X ∈ Rn . As already mentioned, this map is linear, and one has Ker(LA ) = {X ∈
Rn | AX = 0}, i.e. Ker(LA ) are the solutions of the linear system AX = 0.

Remark 4.3.3. The kernel of a linear map is never empty, indeed it always contains
the element 0.
66 CHAPTER 4. LINEAR MAPS

Lemma 4.3.4. Let T : V → W be a linear map between vector spaces over the same
field F, and assume that Ker(T) = {0}. If {X1 , . . . , Xr } are linearly independent ele-
ments of V , then {T(X1 ), . . . , T(Xn )} are linearly independent elements of W .

Proof. Let λ1 , . . . , λn such that

λ1 T(X1 ) + λ2 T(X2 ) + · · · + λn T(Xn ) = 0.


( )
By linearity, this is equivalent to T λ1 X1 + · · · + λn Xn = 0, but since the kernel of T is
reduced to 0 it means that λ1 X1 + · · · + λn Xn = 0. Finally, by the linear independence
of X1 , . . . , Xn it follows that λj = 0 for any j ∈ {1, . . . , n}. As a consequence, the
elements T(X1 ), . . . , T(Xn ) of W are linearly independent.

Let us now come to an important result of this section. For this, we just recall that
for a vector space, its dimension corresponds to the number of elements of any of its
bases. It also corresponds to the maximal number of linearly independent elements of
this vector space.

Theorem 4.3.5. Let T : V → W be a linear map between two vector spaces over the
same field F, and assume that V is of finite dimension. Then
( ) ( )
dim Ker(T) + dim Ran(T) = dim(V ).

Proof. Let {Y1 , . . . , Yn } be a basis for Ran(T), and let X1 , . . . , Xn ∈ V such that
T(Xj ) = Yj for any j ∈ {1, . . . , n}. Let also {K1 , . . . , Km } be a basis for Ker(T). Note
that if one shows that {X1 , . . . , Xn , K1 , . . . , Km } is a basis for V , then the statement is
proved (with dim(V ) = m + n).
So, let X be an arbitrary element of V . Then there exist λ1 , . . . , λn ∈ F such that
T(X) = λ1 Y1 + · · · + λn Xn , since {Y1 , . . . , Yn } is a basis for Ran(T). It follows that

0 = T(X) − λ1 X1 − · · · − λn Yn
= T(X) − λ1 T(X1 ) − · · · − λn T(Xn )
( )
= T X − λ1 X1 − · · · − λn Xn ,

which means that X − λ1 X1 − · · · − λn Xn belongs to Ker(T). As a consequence, there


exist λ′1 , . . . , λ′m ∈ F such that

X − λ1 X1 − · · · − λn Xn = λ′1 K1 + · · · + λ′m Km ,

since {K1 , . . . , Km } is a basis for Ker(T). Consequently, one gets

X = λ1 X1 + · · · + λn Xn + λ′1 K1 + · · · + λ′m Km ,

or in other words ( )
Vect X1 , . . . , Xn , K1 , . . . , Km = V.
4.4. RANK AND LINEAR MAPS 67

Let us now show that these vectors are linearly independent. By contraposition,
assume that
λ1 X1 + · · · + λn Xn + λ′1 K1 + · · · + λ′m Km = 0 (4.3.1)
for some λ1 , . . . , λn , λ′1 , . . . , λ′m . Then one infers from (4.3.1) that

0 = T(0)
( )
= T λ1 X1 + · · · + λn Xn + λ′1 K1 + · · · + λ′m Km
( )
= T λ1 X1 + · · · + λn Xn + 0
= λ1 T(X1 ) + · · · + λn T(Xn )
= λ1 Y1 + · · · + λn Yn .

Since Y1 , . . . , Yn are linearly independent, one already concludes that λj = 0 for any
j ∈ {1, . . . , n}. It then follows from (4.3.1) that λ′1 K1 + · · · + λ′m Km = 0, which implies
that λ′i = 0 for any i ∈ {1, . . . , m} since the vectors Ki are linearly independent.
In summary, one has shown that V is generated by the family of linearly independent
elements X1 , . . . , Xn , K1 , . . . , Km of V . Thus, these elements define a basis, as expected.

4.4 Rank and linear maps


Let us come back to matrices over F. For any A ∈ Mmn (F), recall that we denote by
Aj the j th column of A and by Ak the k th row of A. We also denote by LA : Fn → Fm
the linear map defined by LA (X) = AX. Observe finally that {Ej }nj=1 is a basis of Fn
(note that the 1 at the entry j of Ej is the 1 of the field F). Thus, for any X ∈ Fn one
has
X = t (x1 , . . . , xn ) = x1 E1 + x2 E2 + · · · + xn En
and in addition
( )
LA (X) = A x1 E1 + x2 E2 + · · · + xn En
= x1 AE1 + x2 AE2 + · · · + xn AEn
= x1 A1 + x2 A2 + . . . xn An .

With such equalities, one directly infers the following statement:

Lemma 4.4.1. The range of LA corresponds to the subspace generated by the columns
of A.

Proof. It is enough to remember the following equality


{ }
Ran(LA ) = LA (X) | X ∈ Fn

and to take into account the computation performed before the statement.
68 CHAPTER 4. LINEAR MAPS

Considering the dimensions of these spaces one directly gets:


Corollary 4.4.2. The dimension of the range of the linear map LA is equal to the rank
of A, i.e. ( )
dim Ran(LA ) = rank(A).
( )
Theorem 4.4.3. Let A ∈ Mmn (F) with rank(A) = r. Then one has dim Ker(LA ) =
n − r.
Proof. Since LA : Fn → Fm is a linear map, one has from Theorem 4.3.5
( ) ( )
dim Ker(LA ) + dim Ran(LA ) = n,
| {z }
=r

from which the statement follows.


Example 4.4.4. What is the dimension of the space of solutions of the system
{
2x1 − x2 + x3 + 2x4 = 0
?
x1 + x2 − 2x3 − x4 = 0
(x )
1 ( )
Since this system is equivalent to LA xx23 = ( 00 ) with A = 21 −1 1 2
1 −2 −1 and since
x4
( )
rank(A) = 2, one directly infers from the previous result that dim Ker(LA ) = 4−2 = 2.
This corresponds to the dimension of the space of solutions of the homogeneous equation.
One ends up this section with an important result:
Theorem 4.4.5. Let A ∈ Mmn (F) and B ∈ Fm , and consider the equation AX = B
for some X ∈ Fn . If this
( equation
) has a solution, then its set of all solutions is of
dimension equal to dim Ker(LA ) .
Proof. Assume that Y0 ∈ Fn satisfies AY0 = B. Then if Y ∈ Fn satisfies AY = 0,
one infers that A(Y0 + Y ) = B, which means that Y0 + Y is a solution of the original
problem, for any Y ∈ Ker(LA ). Now, if one can show that all solutions of AX = B
are of the form X = Y0 + Y for some Y ∈ Ker(LA ), then the statement is proved. For
that purpose, it is sufficient to observe that if X ∈ Fn satisfies AX = B, then one has
A(X − Y0 ) = B − B = 0, or in other words X − Y0 =: Y for some Y ∈ Ker(LA ). As a
consequence, one infers that X = Y0 + Y with Y ∈ Ker(LA ), as expected.

4.5 Matrix associated with a linear map


Let us start with a question: If V, W are vector spaces over a field F and if T : V → W
is a linear map, how can one associate with this linear map a matrix ?
In fact, this can be done only once a choice of bases for V and W has been done,
and the resulting matrix will depend on the choice of bases, as we shall see. So, let
4.5. MATRIX ASSOCIATED WITH A LINEAR MAP 69

us introduce a new notation: a basis for a vector space V over F will be denoted by
V := {V1 , . . . , Vn } with {V1 , . . . , Vn } a family of linearly independent elements of V
which generate V . In addition, let us denote by X an arbitrary element of V (which
was simply denoted by X up to now). Then, since V is a basis for V there exists
X := t (x1 , . . . , xn ) ∈ Fn such that

X = x1 V1 + x2 V2 + · · · + xn Vn .

The vector X ∈ Fn is called the coordinate vector of X with respect to the basis V of V ,
and we shall use the notation
(X )V = X
meaning precisely that the coordinates of X with respect to the basis V are X.
Remark 4.5.1. Clearly, if V = Rn and if Vj = Ej , one just says that X are the
coordinates of X and one uses to identify X and X. This is what we have done until
now since we have only considered the usual basis {Ej }nj=1 on Rn . However, if one needs
to consider different bases on Rn , the above notations are necessary. Note for example
that X exists without any choice of a particular basis, while X depends on such a choice.
Now, if Y is another element of V with (Y)V = Y = t (y1 , . . . , yn ), let us observe
that
(X + Y)V = X + Y and (λX )V = λX (4.5.1)
for any λ ∈ F. Indeed, this follows from the equalities

X + Y = x1 V1 + · · · + xn Vn + y1 V1 + · · · + yn Vn
= (x1 + y1 )V1 + · · · + (xn + yn )Vn

and
λX = λ(x1 V1 + · · · + xn Vn ) = (λx1 )V1 + · · · + (λxn )Vn .
Thus, choosing a basis V for V allows one to identity any point of V with an element
of Fn via its coordinate vector. By taking (4.5.1) into account, one also observes that
V allows one to define a linear map (·)V : V → Fn .
We also consider a vector space W over F endowed with a basis W := {W1 , . . . , Wm }.
In this case, for any Z ∈ W we set (Z)W = Z = t (z1 , . . . , zm ) ∈ Fm for the coordinate
vector of Z with respect to the basis W of W . Thus, if T : V → W is a linear map,
there exists T := (tij ) ∈ Mmn (F), called the matrix associated with T with respect to
the basis V of V and W of W defined by

m ∑
m
t
T(Vj ) = tij Wi = tji Wi (4.5.2)
i=1 i=1

for any j ∈ {1, . . . , n}. On the other hand, we shall show just below that the following
equality also holds ( )
T(X ) W = T (X )V . (4.5.3)
70 CHAPTER 4. LINEAR MAPS

In other words, the action of T on a basis of V is given in terms of t T by relation


(4.5.2), while the action of T on the coordinate vectors is given in terms of T by
relation (4.5.3). Note that this is related to the more general notion of covariant or
contravariant transformations.
For the proof of (4.5.3) it is enough to observe that one has
(∑
n ) ∑
n
T(X ) = T xj Vj = xj T(Vj )
j=1 j=1
∑n ∑
m m (∑
∑ n ) ∑
m
= xj tij Wi = tij xj Wi = (T X)i Wi ,
j=1 i=1 i=1 j=1 i=1

which implies that ( )


T(X ) W = T X = T (X )V . (4.5.4)
Example 4.5.2. If V = Rn , W = Rm and Vj = Ej while Wi = Ei for any j ∈
{1, . . . , n} and j ∈ {1, . . . , m}, and if T is a linear map from Rn to Rm then one
deduces from (4.5.2) that
 
t1j
∑m  t2j 
 
T(Ej ) = tij Ei =  ..  = T j
i=1
 . 
tmj

where T j corresponds to the j th column of the matrix T . In other words one has
( )
T = T(E1 ) T(E2 ) . . . T(En ) .

Example 4.5.3. If V is a real vector space with basis V = {V1 , V2 , V3 } and if T : V → V


is the linear map such that

T(V1 ) = 2V1 − V2 , T(V2 ) = V1 + V2 − 4V3 , T(V3 ) = 5V1 + 4V2 + 2V3 ,

then the matrix associated with T with respect to the basis V is given by
 
2 1 5

T = −1 1 4 .
0 −4 2
Let us still consider the notion of a change of basis. Indeed, given the matrix associ-
ated to a linear map in a prescribed basis, it is natural to wonder about the matrix asso-
ciated to the same linear map but with respect to another basis. So, let V = {V1 , . . . , Vn }
and V ′ = {V1′ , . . . , Vn′ } be two basis of the same vector space V . Let B = (bij ) ∈ Mn (F)
be the matrix defined by
∑n ∑
n

Vj = bij Vi ≡ t
bji Vi .
i=1 i=1
4.6. COMPOSITION OF LINEAR MAPS 71

It is easily observed that the matrix B is invertible. Then, for any X ∈ V with X = (X )V
and X ′ = (X )V ′ , one has


n ∑
n ∑
n ∑
n n (∑
∑ n )
xi Vi = X = x′j Vj′ = x′j bij Vi = bij x′j Vi .
i=1 j=1 j=1 i=1 i=1 j=1

Since the vectors V1 , . . . , Vn are linearly independent, this implies that

X = BX ′ or equivalently (X )V = B(X )V ′ . (4.5.5)

Let us now consider a linear map T : V → V , and let T be the matrix associated
with T with respect to the basis V, and let T ′ be the matrix associated to T with
respect to the basis V ′ . The original question corresponds then to the link between T
and T ′ ? In order to answer this question, observe that for any X ∈ V one gets by
equations (4.5.4) and (4.5.5) that
( ) ( )
T BX ′ = T X = T(X ) V = B T(X ) V ′ = BT ′ X ′ .

Since X ′ is arbitrary, one infers that T B = BT ′ , or equivalently

T ′ = B −1 T B. (4.5.6)

One deduces in particular that the matrix T and T ′ are similar, see Definition 2.1.16.
Note that a similar (but slightly more complicated) computation can be realized
for a linear map between two vector spaces V and W over the same field F endowed
with two different bases V, V ′ and W, W ′ .

4.6 Composition of linear maps


Let us now consider three sets U, V, W and let F : U → V and G : V → W be maps.
Then the map
G ◦ F : U → W,
( ) ( )
defined by G ◦ F (X) = G F(X) for any X ∈ U , is called the composition map of F
with G. Notice that if W ̸⊂ U the composition map F ◦ G has simply no meaning.

Examples 4.6.1. (i) Let U = V = W = R and F, G be two real functions defined


on R. Then G ◦ F just corresponds to the composition of functions.

(ii) If U = Rn , V = Rm , W = Rp , A ∈ Mmn (R) and B ∈ Mpm (R), then for any


X ∈ Rn one has
( ) ( )
LB ◦ LA (X) = LB LA (X) = BAX = (BA)X = LBA (X). (4.6.1)
72 CHAPTER 4. LINEAR MAPS

Let us now observe an important property of the composition of maps, namely the
associativity. Indeed, If U, V, W, S are sets and F : U → V , G : V → W and H : W → S
are maps, one has
(H ◦ G) ◦ F = H ◦ (G ◦ F).
Indeed, for any X ∈ U one has
[ ] ( ) ( )
(H ◦ G) ◦ F (X) = H ◦ G (F(X)) = H G(F(X))

and [ ] ( ) ( )
H ◦ (G ◦ F) (X) = H (G ◦ F)(X) = H G(F(X)) ,
and the equality of the two right hand sides implies the statement.
Lemma 4.6.2. Let U, V, W be vector spaces over a field F, and let G : U → V ,
G′ : U → V , H : V → W and H′ : V → W be linear maps. Then
(i) H ◦ G : U → W is a linear map,

(ii) (H + H′ ) ◦ G = H ◦ G + H′ ◦ G,

(iii) H ◦ (G + G′ ) = H ◦ G + H ◦ G′ ,

(iv) (λH) ◦ G = H ◦ (λG) = λ(H ◦ G), for all λ ∈ F.


The proof will be provided in Exercise 4.17.
Remark 4.6.3. If V is a vector space and if T : V → V is a linear map, then Tn =
| ◦ T{z
T · · · ◦ T} is a linear map from V to V . By convention, one sets T0 = 1, and observes
n terms
that Tr+s = Tr ◦ Ts = Ts ◦ Tr .

4.7 Inverse of a linear map


Definition 4.7.1. For a map F : V → W between two sets V and W , one says that F
has an inverse if there exists G : W → V such that G ◦ F = 1V and F ◦ G = 1W . In
this case, one also says that F is invertible and write F−1 for this inverse.
Example 4.7.2. If A ∈ Mn (R) is invertible, then the linear map LA : Rn → Rn is
invertible, with inverse LA−1 . This follows from equation (4.6.1), or more precisely

LA ◦ LA−1 = LAA−1 = 1n = LA−1 A = LA−1 ◦ LA .

Due to the following lemma, there is no ambiguity in speaking about the inverse
(and not only about an inverse) of a invertible map.
Lemma 4.7.3. Let F : V → W be an invertible map between two sets V et W . Then
this inverse is unique.
4.7. INVERSE OF A LINEAR MAP 73

Proof. Let us assume that there exists G : W → V and G′ : W → V such that


G ◦ F = 1V , F ◦ G = 1W , G′ ◦ F = 1V , and F ◦ G′ = 1W . Then one gets

G = 1V ◦ G = (G′ ◦ F) ◦ G = G′ ◦ (F ◦ G) = G′ ◦ 1W = G′

from which the result follows.

Let us now come to two more refined notions related to a maps, linear or not.

Definition 4.7.4. A map F : V → W between two sets is injective or one-to-one if


F(X1 ) ̸= F(X2 ) whenever X1 , X2 ∈ V with X1 ̸= X2 . The map F is called surjective
if for any Y ∈ W there exists at least one X ∈ V such that F(X) = Y . The map F is
bijective if it is both injective and surjective.

The following result links the notions of invertibility and of bijectivity.

Theorem 4.7.5. A map F : V → W between two sets is invertible if and only if F is


bijective.

Proof. (i) Assume first that F is bijective. In particular, since F is surjective, for any
Y ∈ W , there exists X ∈ V such that F(X) = Y . Note that X is unique because F is
also injective. Thus if one sets F−1 (Y ) := X then one has
) ( )
(F−1 ◦ F (X) = F−1 F(X) = F−1 (Y ) = X

which implies that F−1 ◦ F = 1V , and similarly


( ) ( )
F ◦ F−1 (Y ) = F F−1 (Y ) = F(X) = Y

which implies that F ◦ F−1 = 1W . One has thus define an inverse for F.
(ii) Let us now assume that F is invertible, with inverse denoted by F−1 . Let first
X1 , X2 ∈ V with F(X1 ) = F(X2 ). One then deduces that
( ) ( ) ( ) ( )
X1 = 1V X1 = F−1 ◦ F X1 = F−1 F(X1 ) = F−1 F(X2 ) = F−1 ◦ F (X2 ) = X2 ,

and thus F is injective. Secondly, let Y ∈ W , and observe that


( ) ( )
Y = 1W Y = F ◦ F−1 (Y ) = F F−1 (Y )

which implies that Y = F(X) for X given by F−1 (X). Thus F is injective. Since F is
both injective and surjective, F is bijective.

For linear maps the general theory simplifies a lot, as we shall see now.

Theorem 4.7.6. Let V, W be two vector spaces over the same field F, and let T : V →
W be an invertible linear map. Then its inverse T−1 : W → V is also a linear map.
74 CHAPTER 4. LINEAR MAPS

Proof. Let Y1 , Y2 ∈ W and set X1 := T−1 (Y1 ) and X2 := T−1 (X2 ). Since T ◦ T−1 = 1W
one has for j ∈ {1, 2}
( ) ( )
Yj = T ◦ T−1 (Yj ) = T T−1 (Yj ) = T(Xj ).

Then, one infers that


( ) ( ) ( )
T−1 Y1 + Y2 = T−1 T(X1 ) + T(X2 ) |{z}
= T−1 T(X1 + X2 )
linearity
−1 −1
=T ◦ T(X1 + X2 ) = X1 + X2 = T (Y1 ) + T−1 (Y2 ). (4.7.1)

Similarly one has for any λ ∈ F and Y ∈ W (with Y := T(X))


( ) ( )
T−1 (λY ) = T−1 λT(X) |{z}
= T−1 T(λX)
linearity
( −1
)
= T ◦ T (λX) = λX = λT−1 (Y ). (4.7.2)

It is then sufficient to observe that (4.7.1) and (4.7.2) correspond to the linearity con-
ditions for T−1 .
In the next statement we give an equivalent property for the injectivity of a linear
map.
Lemma 4.7.7. A linear map T : V → W between two vector spaces over the same field
is injective if and only if Ker(T) = {0}.
Proof. (i) The first part of the proof is a contraposition argument: instead of proving
A ⇒ B we show equivalently that B̄ ⇒ Ā. Thus, let us assume first that Ker(T) ̸= {0},
then there exists X0 ̸= 0 such that T(X0 ) = 0. In addition, for any X ∈ V one has

T(X + X0 ) = T(X) + T(X0 ) = T(X) + 0 = T(X).

Since X ̸= X + X0 but T(X) = T(X + X0 ), one concludes that T is not injective. By


contraposition, one has shown that T injective implies that Ker(T) = {0}.
(ii) Assume now that Ker(T) = {0}, and consider X1 , X2 ∈ V with X1 ̸= X2 . Then
one has
T(X1 ) − T(X2 ) = T(X1 − X2 ) ̸= 0
since X1 − X2 ̸= 0. As a consequence, T(X1 ) ̸= T(X2 ).
Let us provide a final theorem for this section, which is useful in many situations.
Theorem 4.7.8. Let T : V → W be a linear map between the vector spaces V and W ,
and assume that dim(V ) = dim(W ) < ∞. Then the following assertions are equivalent:
(i) Ker(T) = {0},

(ii) T is invertible,
4.7. INVERSE OF A LINEAR MAP 75

(iii) T is surjective.

Proof. The implication (ii) ⇒ (i) and (ii) ⇒ (iii) are direct consequences of Theorem
4.7.5 and Lemma 4.7.7.
Assume now (i), and recall from Lemma 4.7.7 that this condition corresponds to T
injective. Then from Theorem 4.3.5 and more precisely from the equality
( ) ( )
dim Ker(T) +dim Ran(T) = dim(V )
| {z }
0
( )
one deduces that dim Ran(T) = dim(V ) = dim(W ), where the assumption ( about )the
dimension has been taken into account. It is enough then to observe that dim Ran(T) =
dim(W ) means that T is surjective. Since T is also injective, it follows that T is bijec-
tive. Since bijectivity corresponds to invertibility by Theorem 4.7.5, one infers that (ii)
holds.
Assume now that (iii) holds. By taking again Theorem 4.3.5 into account, one
deduces that from the equality
( )
dim Ran(T) = dim(W ) = dim(V )
( )
that dim Ker(T) = 0, meaning that T is injective. Again, it implies that T is bijective,
and thus invertible, and thus that (ii) holds.
76 CHAPTER 4. LINEAR MAPS

4.8 Exercises
( )
Exercise 4.1. Let F : R2 → R2 be the map defined by F ( xy ) = 2x 3y for any ( xy{) ∈ R2 .
Describe the image
} by F of the points lying on the unit circle centered at 0, i.e. ( xy ) ∈
R2 | x2 + y 2 = 1 .

y ) for any ( y ) ∈ R .
Exercise 4.2. Let F : R2 → R2 be the map defined by F ( xy ) = ( xy x 2

Describe the image by F of the line {( y ) ∈ R | x = 2}.


x 2

Exercise 4.3. Let V be a vector space of dimension n, and let {X1 , . . . , Xn } be a basis
for V . Let F be a linear map from V into itself. Show that F is uniquely defined if one
knows F(Xj ) for j ∈ {1, . . . , n}. Is it also true if F is an arbitrary map from V into
itself ?
Exercise 4.4. Let V, W be vector spaces over the same field, and let T : V → W be a
linear map. Show that the following set is a subspace of V :

{X ∈ V | T(X) = 0}.

This subspace is called the kernel of T.


Exercise 4.5. Show that the image of a convex set under a linear map is a convex set.
Exercise 4.6. Determine which of the following maps are linear:
(x)
a) F : R3 → R2 defined by F yz = ( xz ),

b) F : R4 → R4 defined by F(X) = −X for all X ∈ R4 ,


( 0 )
c) F : R3 → R3 defined by F(X) = X + −1 for all X ∈ R3 ,
0
( )
d) F : R2 → R2 defined by F ( xy ) = 2x
y−x ,

e) F : R2 → R2 defined by F ( xy ) = ( xy ),

f ) F : R2 → R defined by F ( xy ) = xy.
Exercise 4.7. Determine the kernel and the range of the maps defined in the previous
exercise.
Exercise 4.8. Consider the subset of Rn consisting of all vectors t (x1 , . . . , xn ) such that
x1 + x2 + · · · + xn = 0. Is it a subspace of Rn ? If so, what is its dimension ?
Exercise 4.9. Let P : Mn (R) → Mn (R) be the map defined for any A ∈ Mn (R) by
1( )
P(A) = A + tA .
2
1. Show that P is a linear map.
4.8. EXERCISES 77

2. Show that the kernel of P consists in the vector space of all skew-symmetric ma-
trices.

3. Show that the range of P consists in the vector space of all symmetric matrices.

4. What is the dimension of the vector space of all symmetric matrices, and the
dimension of the vector space of all skew-symmetric matrices ?

Exercise 4.10. Let C ∞ (R) be the vector space of all real functions on R which admit
derivatives of all orders. Let D : C ∞ (R) → C ∞ (R) be the map which associates to any
f ∈ C ∞ (R) its derivative, i.e. Df = f ′ .

1. Is D a linear map ?

2. What is the kernel of D ?

3. What is the kernel of Dn , for any n ∈ N, and what is the dimension of this vector
space ?

Exercise 4.11. Consider the map F : R3 → R4 defined by


 
  x
x  x−y 
F y = 

 x − z .

z
x−y−z

1. Is F a linear map ? (Justify your answer)

2. Determine the kernel of F.

3. Determine the range of F.

Exercise 4.12. What is the dimension of the space of solutions of the following systems
of linear equations ? In each case, find a basis for the space of solutions.
{ {
2x + y − z = 0 { 4x + 7y − πz = 0
a) b) x − y + z = 0 c)
2x + y + z = 0 2x − y + z = 0

and 
x + y + z = 0
d) x−y = 0

y+z = 0
( 0 1 3 −2 )
Exercise 4.13. Let A be the matrix given by A = 2 1 −4 3 and consider the linear
2 3 2 −1
map LA : R4 → R3 defined by LA X = AX for all X ∈ R4 .

1. Determine the rank of A and the dimension of the range of LA .


78 CHAPTER 4. LINEAR MAPS

2. Deduce the dimension of the kernel of LA , and exhibit a basis for the kernel of LA .
( )
0
3. Find the set of all solutions of the equation AX = 2 .
2

Exercise 4.14. Let F : R3 → R2 be the map indicated below. What is the matrix
associated with F in the canonical bases of R3 and R2 ?
( ) ( ) ( )
1 −4 3
a) F(E1 ) = , F(E2 ) = , F(E3 ) =
−3 2 1

and  
x1 ( )
3x − 2x + x
b) F x2  = 1 2 3
.
4x1 − x2 + 5x3
x3

) 4.15. Let L : R → R be a linear map which associated matrix has the form
3 3
Exercise
(
1 0 0
0 2 0 with respect to the canonical basis of R3 . What is the matrix associated with L
0 0 3 ( √ ) ( √ ) ( 0 )
1/ 2 −1/ 2
in the basis generated by the three vectors V1 = 1/ 2 , V2 = 1/√2 , V3 = −1

0
0 0

Exercise 4.16. For any A, B ∈ Mn (R), one says that A and B commute if AB = BA.

a) Show that the set of all matrices which commute with A is a subspace of Mn (R),

b) If A = ( 02 13 ), exhibit a basis of the subspace of all matrices which commute with


A.

Exercise 4.17. Let U, V, W be vector spaces over a field F, and let G : U → V ,


G′ : U → V , H : V → W and H′ : V → W be linear maps. Show that

(i) H ◦ G : U → W is a linear map,

(ii) (H + H′ ) ◦ G = H ◦ G + H′ ◦ G,

(iii) H ◦ (G + G′ ) = H ◦ G + H ◦ G′ ,

(iv) (λH) ◦ G = H ◦ (λG) = λ(H ◦ G), for all λ ∈ F.

Exercise 4.18. Let V be a real vector space, and let P : V → V be a linear map
satisfying P2 = P. Such a linear map is called a projection.

(i) Show that 1 − P is also a projection, and that (1 − P)P = P(1 − P) = 0,

(ii) Show that V = Ker(P) + Ran(P),

(iii) Show that the intersection of KerP and Ran(P) is {0}.


4.8. EXERCISES 79
( x+y )
Exercise 4.19. Let L : R2 → R2 be the linear map defined by L ( xy ) = x−y . Show that
L (is invertible
) ( x−y and)find its inverse. Same question with the map L : R → R defined by
3 3
x
L yz = x+z .
x+y+3z

Exercise 4.20. Let F, G be invertible linear maps from a vector space into itself. Show
that (G ◦ F)−1 = F−1 ◦ G−1 .
Exercise 4.21. Show that the matrix B : Rn → Rn defining a change of basis in Rn is
always invertible.
Exercise 4.22. Let V be the set of all infinite sequences of real numbers (x1 , x2 , x3 , . . . ).
We endow V with the pointwise addition and multiplication, i.e.

(x1 , x2 , x3 , . . . ) + (x′1 , x′2 , x′3 , . . . ) = (x1 + x′1 , x2 + x′2 , x3 + x′3 , . . . )

and λ(x1 , x2 , x3 , . . . ) = (λx1 , λx2 , λx3 , . . . ), which make V an infinite dimensional vec-
tor space.
Define the map F : V → V , called shift operator, by

F(x1 , x2 , x3 , . . . ) = (0, x1 , x2 , x3 , . . . ).

(i) Is F a linear map ?

(ii) Is F injective, and what is the kernel of F ?

(iii) Is F surjective ?

(iv) Show that there is a linear map G : V → V such that G ◦ F = 1.

(v) Does the map G have the property that F ◦ G = 1 ?

(vi) What is different from the finite dimensional case, i.e. when V is of finite dimen-
sion ?
Exercise 4.23. Consider the matrices
( ) ( ) ( ) ( )
2 0 1 0 −1 0 0 1
A= , B= , C= , D= ,
0 2 0 0 0 1 −1 0
( ) ( )
1 0.2 1 −1
E= , F= ,
0 1 1 1

and show their effect on the letter L defined by the three points ( 02 ) , ( 00 ) , ( 10 ) of R2 .
Exercise 4.24. Let N = ( nn12 ) be a vector in R2 with ∥N ∥ = 1, and let ℓ be the line in
R2 passing trough 0 ∈ R2 and parallel to N . Then any vector X ∈ R2 can be written
uniquely as X = X∥ + X⊥ , where X∥ is a vector parallel to ℓ and X⊥ is a vector
perpendicular to ℓ. Show that there exists a projection P ∈ M2 (R) such that X∥ = P X,
and express P in terms of n1 and n2 .
80 CHAPTER 4. LINEAR MAPS
( n1 )
Exercise 4.25. 1) Do the same exercise in R3 with N given by nn23 .
2) Show that there also exists a projection Q such that X⊥ = Q X. If H0,N is the
plane passing through 0 ∈ R3 and perpendicular to N , show that X⊥ ∈ H0,N .

Exercise 4.26. In the framework of the previous exercise, a reflection of X about H0,N
is defined by the vector Xref := X⊥ − X∥ . Show that ∥Xref ∥ = ∥X∥, and provide the
expression for the linear map transforming X into Xref .

Exercise 4.27. Block matrices are matrices which are partitioned into rectangular sub-
matrices called blocks. For example, let A ∈ Mn+m (R) be the block matrix
( )
A11 A12
A=
A21 A22

with A11 ∈ Mn (R), A22 (R) ∈ Mm (R), A12 ∈ Mn×m (R), and A21 ∈ Mm×n (R). Such
matrices can be multiplied as if every blocks where scalars (with the usual multiplication
of matrices), as long as the products are well defined. For example, check this statement
by
( computing ) the product 0AB in two different ways with the
−1
( B11following
B12
) matrices: A =
A11 A12 with A11 = ( 1 0 ) and A12 = ( 1 ), and B = B21 B22 with B11 = ( 14 25 ),
1
( ) ( )
B12 = ( 36 ), B21 = 7 8 , and B22 = 9 .

Exercise 4.28. Let A ∈ Mn+m (R) be the block matrix


( )
A11 A12
A=
O A22

with A11 ∈ Mn (R), A22 (R) ∈ Mm (R) and A12 ∈ Mn×m (R).

(i) For which choice of A11 , A12 and A22 is A invertible ?

(ii) If A is invertible, what is A−1 , in terms of A11 , A12 and A22 ?

You might also like