Further Mathematical Methods (Linear Algebra) Solutions For Problem Sheet 2
Further Mathematical Methods (Linear Algebra) Solutions For Problem Sheet 2
Further Mathematical Methods (Linear Algebra) Solutions For Problem Sheet 2
1. We are asked to prove Theorems 3.3 and 3.7 from the hand-out, namely:
a. T (0) = 0.
b. T (v) = T (v).
c. T (u v) = T (u) T (v).
T (1 u + 2 v) = 1 T (u) + 2 T (v),
where T (u), T (v) W . The results that we have to prove follow directly from this
definition, i.e. for any vectors u, v V :
Thus, T (0) = 0.
b. Setting 1 = 0 and 2 = 1 in the definition we get
as required.
and,
Proof: We are given that V and W are vector spaces and that T : V W is a linear
transformation. We prove the theorem by taking each case in turn:
1
To show that the null space of T , denoted by N (T ), is a subspace of V we note that
N (T ) = {v V | T (v) = 0},
T (x + y) = T (x) + T (y) = 0 + 0 = 0,
and so x + y N (T ).
Closure under scalar multiplication since for any scalar :
T (x) = T (x) = 0 = 0,
and so x N (T ).
Thus, N (T ) is a subspace of V , as required.
To show that the range of T , denoted by R(T ), is a subspace of W we note that
and so x+y R(T ) as there is a vector x0 +y0 V such that T (x0 +y0 ) = x+y.1
Closure under scalar multiplication since for any scalar :
T (x0 ) = T (x0 ) = x,
T (x + 0 x0 ) = T (x) + 0 T (x0 ),
for all vectors x, x0 R4 and all scalars , 0 R. To do this, we take two general vectors x =
[w, x, y, z]t and x0 = [w0 , x0 , y 0 , z 0 ]t in R4 and any two real numbers and 0 and note that
0
w w w + 0 w0
x 0 0 0
T (x + 0 x0 ) = T + 0 x = T x + x ,
y y
0 y + 0 y 0
z z0 z + 0 z 0
1
Notice that x0 + y0 V since x0 , y0 V and V is a vector space (i.e. it is closed under vector addition itself).
2
Notice that x0 V since x0 V and V is a vector space (i.e. it is closed under scalar multiplication itself).
3
That is, Definition 3.1.
2
which on applying the transformation yields
0
0 0 (w + 0 w0 ) + (x + 0 x0 ) + (y + 0 y 0 ) w+x+y 0 w +x +y
0 0
T (x + x ) = = + .
(x + 0 x0 ) + (y + 0 y 0 ) + (z + 0 z 0 ) x+y+z x0 + y 0 + z 0
where v = [w, x, y, z]t is any vector in R4 . As such, a general vector in R(T ) can be written as
1 1 1 0
T (v) = w +x +y +z ,
0 1 1 1
for any w, x, y, z R. Thus, noting the linear dependence, we can see that
1 0
R(T ) = Lin , = R2 ,
0 1
N (T ) = {v R4 | T (v) = 0},
is just the set of all vectors of the form [w, x, y, z]t whose components satisfy the vector equation
w+x+y
= 0.
x+y+z
So, equating components, we are looking for vectors in R4 whose components satisfy the simul-
taneous equations:
w+x+y =0
x+y+z =0
Now, as we have two equations in four variables, we take x and y (say) to be the free parameters
r and s respectively. Thus, w = rs and z = rs, which means that the null space contains
vectors of the form
w 1 1
x 1 0
= r + s ,
y 0 1
z 1 1
3
and so, as these vectors are linearly independent, we have
1 1
1 , 0 ,
N (T ) = Lin 0 1
1 1
We are also asked to verify the rank-nullity theorem which states that
(T ) + (T ) = n,
where (T ) is the dimension of the range of T (i.e. 2), (T ) is the dimension of the null space of
T (i.e. 2) and n is the dimension of the domain of the transformation (i.e. 4). Comparing these
numbers, it is easy to verify that the rank-nullity theorem holds in this case (as 2 + 2 = 4).
Further, we are asked to find the matrix for T with respect to the standard bases in R4 and R2 .
To do this, we note that the standard basis of R4 is given by the set of vectors
1 0 0 0
0 1 0
, , , 0 ,
0 0 1 0
0 0 0 1
and so calculating the image of each of these vectors under the transformation T we find that
1 0 0 0
0 1 1 1 0 1 0 0
T =
, T =
, T =
, and T = ,
0 0 0 1 1 1 0 1
0 0 0 1
where the vectors on the left-hand-sides of these expressions are automatically of the required form
as we are dealing with the standard basis of R2 too. Thus, in this case,
1 1 1 0
AT = .
0 1 1 1
And, we are also asked to find the the matrix for T with respect to the bases
1 0 1 0
2 1 0 0 0 1 1
S = , , , and S = , ,
0 0 1 2 1 2
1 1 1 1
of R4 and R2 respectively. So, as before, we start by calculating the image of each of these vectors
under the transformation T , i.e.
1 0 1 0
2
T = 3 , T 1 = 1 , T 0 = 2 , and T 0 = 2 ,
0 3 0 2 1 2 2 3
1 1 1 1
where the vectors on the left-hand-sides of these expressions are written in terms of the standard
basis of R2 . However, to find AT in this case, we need them to be written in terms of the basis
S 0 = {w1 , w2 } where w1 = [1, 1]t and w2 = [1, 2]t , i.e. we need to find [T (v)]S 0 for these vectors. To
do this, we denote the vectors in S by v1 , v2 , v3 and v4 , and note that the first expression gives
3 1 1 3
T (v1 ) = =3 +0 = 3w1 + 0w2 = [T (v1 )]S 0 = ,
3 1 2 0 S0
4
the second expression gives
1 1 1 0
T (v2 ) = =0 +1 = 0w1 + 1w2 = [T (v2 )]S 0 = ,
2 1 2 1 S0
Consequently,
3 0 2 1
AT = ,
0 1 0 1
which is non-zero. Thus, S is a linearly independent set and hence a basis. Whilst, for S 0 , we
obviously have a linearly independent set as neither vector is a scalar multiple of the other, and so
S 0 is a basis too.
3. We are asked to use the Wronskian to show that the set of functions {ex , ex , ex } [defined for
all values of x R] is linearly independent if , and are all different. To do this, we note that
the Wronskian of a set of three functions, say {f1 , f2 , f3 }, is given by
f1 f2 f3
W (x) = f10 f20 f30 ,
f 00 f 00 f 00
1 2 3
and we know that these functions are linearly independent if there is some x in their domain such
that W (x) 6= 0. So, in the case under consideration, the Wronskian is given by
x
e ex ex 1 1 1
x
W (x) = e ex ex = e(++)x ,
2 ex 2 ex 2 ex 2 2 2
where we have taken the common factor out of each column in the last step. Now, to show that this
set of functions is linearly independent if , and are all different, it is convenient to simplify the
5
determinant by using some column operations, i.e.
1 1 1 1 0 0
= :C2 C2 C1 and C3 C3 C1 .
2 2 2 2 2 2 2 2
1 0 0
= ( )( ) 1 1 :Common factors from C2 and C3 .
2 + +
1 1 1 1 0 0
= ( )( ) 1 0 :C3 C3 C2 .
2 2 2 2 +
So, evaluating the determinant we find that the Wronskian is given by:
W (x) = ( )( )( )e(++)x ,
which is non-zero if , and are all different.4 Thus, the set of functions {ex , ex , ex } [defined
for all values of x R] is linearly independent in this case.
4. To show that the Wronskian for the functions f (x) and g(x) defined [for all values of x R] by
2
2 x for x 0
f (x) = x and g(x) =
0 for x 0
and so,
If x 0, then 2
x x2
W (x) = = 0,
2x 2x
as the two columns are equal.
If x 0, then 2
x 0
W (x) = = 0,
2x 0
as the there is a column of zeroes.
But, we also note that this pair of functions is linearly independent as there is no [single] scalar R
such that g(x) = f (x) for all x in the domain.
Aside: This result is useful because, from the lectures, we have the theorem:
Let the functions f1 , f2 , . . . , fn be defined and n 1 times differentiable for all values of
x in the interval (a, b) R. If W (x) 6= 0 for some x (a, b), then the set of functions
{f1 , f2 , . . . , fn } is linearly independent.
However, this question gives us a counterexample to [the contrapositive of] the converse of this
theorem, i.e.
4
Note that the e(++)x factor in this expression is never zero. Indeed, we only need to show that the Wronskian
is non-zero for some value of x in the domain of these functions. So, as the domain is specified to be all x R, we
could have set x = 0. Then, working with W (0) = ( )( )( ), the result follows.
6
Let the functions f1 , f2 , . . . , fn be defined and n 1 times differentiable for all values of
x in the interval (a, b) R. If W (x) = 0 for all x (a, b), then the set of functions
{f1 , f2 , . . . , fn } is linearly dependent.
since we have two functions where the Wronskian is equal to zero for all x R and yet the functions
are linearly independent.
5. Writing the set of simultaneous equations given in the question in the form Ax = b we find that
x1
1 1 1 1 4
2 0 1 1 x2 = 2 ,
x3
0 2 1 3 6
x4
To solve this we write this equation as an augmented matrix and perform row operations on it until
we reach the Row Reduced Echelon form, i.e.
1 1 1 1 4 1 1 1 1 4
R2 R2 2R1 R R3 +R2
2 0 1 1 2 0 2 1 3 6 3
R2 R2
0 2 1 3 6 0 2 1 3 6
1 1 1 1 4 2 0 1 1 2
R1 2R1 +R2
0 2 1 3 6 0 2 1 3 6 .
0 0 0 0 0 0 0 0 0 0
7
Moreover, we can see that the set of vectors {u1 , u2 } is a basis for the null space since this space is
the set of all solutions of the matrix equation Ax = 0 and using the analysis above, we know that
this is the same as the set of all solutions of the matrix equation
1 1
x1
1 0 2 2 0
0 1 1 3 x2
2 2 x3 = 0 ,
0 0 0 0 0
x4
where we have set x3 = 1 and x4 = 2 . Thus, writing this vector out in full, we can see that the
null space is given by all linear combinations of the form
x1 1 1
x2 1 1 2 3
x=
x3 = 2
+
2
,
2 0
x4 0 2
i.e. the set of vectors {u1 , u2 } spans the null space of A. Also, since neither of these vectors is a scalar
multiple of the other, they are linearly independent and so, by Definition 2.10, this set of vectors is
a basis of the null space of A as required.
The set of all vectors given by the general solution of our set of simultaneous equations is an affine
set, and following the lectures, the dimension of this affine set is two since it is obtained by shifting
the null space of A (which is two-dimensional) through the fixed vector v. In turn, the Cartesian
equation of this affine set is given by the intersection of the two [three-dimensional] hyperplanes
whose Cartesian equations are
which can be obtained from the first and second components of the general solution by using the fact
that 1 = x3 and 2 = x4 .
Other Problems.
The Other Problems on this sheet were intended to help you revise some properties of matrices which
you should have encountered in MA100. Everyone should have tried these as we will be using these
techniques throughout the course.
You should recall that the row (or column) space of a matrix is defined to be the space spanned by
the row (or column) vectors of a matrix. We shall denote the row and column spaces of a matrix A
by RS(A) and CS(A) respectively.
6. We are asked to use row operations to find the rank of the matrix
1 1 2 1
2 1 2 2
A=
1 2 4 1 .
3 0 0 3
8
This should be revision from MA100, and so performing the appropriate row operations we find that:
1 1 2 1 1 1 2 1
2 1 2 2 0 3 6 0
R
2 R2 2R1 R2R 3
1 2 4 1
R3 R3 +R1 0 1 2 0
R4 R4 3R1
3 0 0 3 0 3 6 0
1 1 2 1 1 0 0 1
0 1 2 0 R1 R1 +R2 0 1 2 0
,
0 3 6 0 R3 R3 3R2 0 0 0 0
R4 R4 3R2
0 3 6 0 0 0 0 0
is the Row Reduced Echelon form of the matrix A. So, since this has two leading ones, the rank of
A is two.
Once we have the Row Reduced Echelon form of A, it is easy to write down a smallest spanning
set for RS(A) and CS(A). To do this, we note that the row space of the matrix A, denoted by RS(A),
is the space spanned by the row vectors of this matrix. So, a smallest spanning set for RS(A) would
be
1 0
, 1 ,
0
0 2
1 0
as the row operations have removed any linear dependence from the row vectors that we started
with.5 Similarly, we note that the column space of the matrix A, denoted by CS(A), is the space
spanned by the column vectors of this matrix. So, as the leading ones in the Row Reduced Echelon
form of A indicate the linear independence of the corresponding column vectors in the original form
of the matrix, we can see that the first two column vectors of the matrix as given in the question
are linearly independent. Thus, a smallest spanning set for CS(A) would be
1 1
2 , 1 .
1 2
3 0
However, the set of vectors which can be found from the first two columns of the Row Reduced
Echelon form of A, namely
1 0
0 , 1 ,
0 0
0 0
is NOT a smallest spanning set for CS(A) since although row operations preserve information about
the dimension of the column space, they destroy information about the specific space that is spanned
by the column vectors. (To see that this is so, notice that these two sets clearly span different spaces,
although these spaces do have the same dimension.)6
and we are asked to find the matrix which represents this transformation with respect to the standard
basis of R4 . To do this, we calculate the image of each of the vectors in the standard basis of R4
5
Also, notice that the dimension of the row space is two, as you should expect. (See also Question 9.)
6
Also, notice that the dimension of the column space is two, as you should expect. (See also Questions 8 and 9.)
9
under the transformation T (see Question 2) and we find that
1 0 0 0
0 1 1 1 0 0 0 0
T = 0 , T = 1 , T = 1 , and T = 0 ,
0 0 1 0
1 0 0 1
0 0 0 1
where the vectors on the left-hand-sides of these expressions are automatically of the required form
because, as no other basis was specified, we use the standard basis of R3 too. Thus, in this case,
1 1 0 0
AT = 0 1 1 0 ,
1 0 0 1
where v = [w, x, y, z]t is any vector in R4 . Multiplying this matrix product out we (unsurpris-
ingly) get
w+x
w = x + y ,
w+z
and as such, a general vector in R(AT ) can be written as
1 1 0 0
w = w 0 + x 1 + y 1 + z 0 ,
1 0 0 1
(notice that this is just the column space of AT ) and so removing any linearly dependent vectors,
we can see that
1 0 0
R(AT ) = Lin 0 , 1 , 0 = R3 ,
1 0 1
and this space is three-dimensional.
N (AT ) = {v R4 | AT v = 0},
10
is just the set of all vectors of the form [w, x, y, z]t whose components satisfy the vector equation
w+x
x + y = 0.
w+z
So, equating components, we are looking for vectors in R4 whose components satisfy the simul-
taneous equations:
w+x=0
x+y =0
w+z =0
Now, as we have three equations in four variables, we take w (say) to be the free parameter r.
Thus, x = r, y = x = r and z = r, which means that the null space contains vectors of
the form
w 1
x 1
= r ,
y 1
z 1
and so we have
1
1
N (AT ) = Lin
1 ,
1
and this subspace of R4 is one-dimensional.
Further, it is easy to verify that the rank-nullity theorem holds for matrices, i.e. that
(AT ) + (AT ) = n,
where (AT ) is the dimension of the range of AT (i.e. 3), (AT ) is the dimension of the null space of
AT (i.e. 1) and n is the dimension of the domain of the transformation (i.e. 4), since 3 + 1 = 4.
The set of all vectors b for which the set of simultaneous equations given by AT x = b is consistent
is just the set of all vectors b for which there is an x R4 that satisfies the matrix equation AT x = b.
That is, it is the set of all vectors b R3 such that AT x = b for some x R4 , which is of course,
the range of AT . Thus, for AT x = b to be consistent, we require that
1 0 0
b R(AT ) = Lin 0 , 1 , 0 = R3 ,
1 0 1
8. We are asked to prove that for any real m n matrix A, R(A) = CS(A). To do this we take the
matrix in question to be
a11 a12 a1n
a21 a22 a2n
A= . .. .. .. ,
.. . . .
am1 am2 amn
and note that the range of A, denoted by R(A), is defined as
11
in this case. Thus, for any vector y R(A), there exists a vector x = [x1 , x2 , . . . , xn ]t Rn such that
y = Ax, i.e.
a a a x x1
11 12 1n 1
a21 a22 a2n x2 x2
m n
R(A) = . .. .. .. .. R .. R
.. . . . . .
am1 am2 amn xn xn
a x + a x + + a x
11 1 12 2 1n n
a21 x1 + a22 x2 + + a2n xn
m
= .. R x , x , . . . , x R
. 1 2 n
am1 x1 + am2 x2 + + amn xn
a11 a12 a1n
a21 a22 a2n
m
= x1 . + x2 . + + xn . R x1 , x2 , . . . , xn R
.. .. ..
am1 am2 amn
a11 a12 a1n
a21 a22 a2n
R(A) = Lin . , . , . . . , .
.. .. ..
am1 am2 amn
and so, R(A) is just the space spanned by the column vectors of A. But, this is the column space of
A, and so we can see that
a11 a12 a1n
a21 a22 a2n
CS(A) = Lin . , . , . . . , . = R(A),
.. .. ..
am1 am2 amn
as required.
9. We are asked to establish that, for any real m n matrix, the dimension of the column space,
the dimension of the row space, and the rank of the matrix are all equal. Firstly, we note that, from
Question 8, R(A) = CS(A) and so
(see Problem Sheet 1, Question 12). Secondly, to establish that the dimension of the row space is
equal to the dimension of the column space, we make the observation8 that
When we find the Row Reduced Echelon form of the matrix A, we locate the leading
ones. Once we have found them, we note that:
The rows which contain them yield linearly independent vectors and we use these
row vectors as a basis for the row space. (Thus, the dimension of the row space is
equal to the number of leading ones.)
The columns which contain them also yield linearly independent vectors and we use
the corresponding column vectors from the original matrix as a basis for the column
space. (Thus, the dimension of the column space is equal to the number of leading
ones.)
(This is one of the major advantages of having the Row Reduced Echelon form of a
matrix.)
8
The proof that this is the case is long and boring (even for me!), and so we appeal to this observation as we only
have to establish that the result holds. Incidentally, the proof proceeds in much the same way except it requires you
to prove some theorems about how row operations affect matrices you should have seen some of these in MA100.
12
and consequently,
dim(CS(A)) = dim(RS(A)).
as required.
Harder Problems.
The Harder Problems on this sheet were intended to give you some further practice in proving results
about vector spaces and linear transformations.
Proof: Let S V be a finite linearly independent set of vectors that is not already a basis
for a finite dimensional vector space V where, say, dim(V ) = n. Clearly, by Definition
2.10, this means that S fails to span V and so we perform the following construction
(note that j 1):
Step 0: Take a vector v0 V such that v0 6 Lin(S) and construct the set S0 =
{v0 } S which will also be linearly independent by Theorem 2.9. If S0 spans V ,
then S0 is a basis for V and we have finished.
Step j: Take a vector vj V such that vj 6 Lin(Sj1 ) and construct the set
Sj = {vj } Sj1 which will also be linearly independent by Theorem 2.9. If Sj
spans V , then Sj is a basis for V and we have finished.
(Notice that the construction above must terminate since, by [the contrapositive of] The-
orem 2.14, we can have no more than n linearly independent vectors in an n-dimensional
space.)9 Thus, S can be augmented to form a new set which is a basis for V by adding
the appropriate vectors, as required.
and secondly,
Proof: Let S V be a finite set of vectors which spans, but is not already a basis for,
a finite dimensional vector space V where, say, dim(V ) = n. Clearly, by Definition 2.10,
this means that S fails to be linearly independent and so, some of the vectors in S must
be expressible as linear combinations of the other vectors in S. We now perform the
following construction (note that j 1):
13
Step j: Take any vector v V which is expressible as a linear combination of the
other vectors in Sj1 and remove it from Sj1 to form the the set Sj 12 which will
also span V .13 If Sj is linearly independent, then Sj is a basis for V and we have
finished.
(Notice that the construction above must terminate since, by [the contrapositive of] the
third theorem in Question 4 of Problem Sheet 1, we need at least n vectors to span an
n-dimensional space.)14 Thus, S can be reduced to form a new set which is a basis for V
by removing the appropriate vectors, as required.
11. We are asked to prove the rank-nullity theorem (i.e. Theorem 3.9 in the hand-out) in the cases
where (T ) = 0 and (T ) = n. We shall consider each of these cases in turn:
Case 1: Assume that (T ) = 0. We further assume (as before) that dim(V ) = n and as such,
we can take the set of vectors S 0 = {v1 , v2 , . . . , vn } to be a basis for V . Now, to prove the
rank-nullity theorem in this case we need to find (T ), and to do this, we want to show that the
set of vectors S 00 = {T (v1 ), T (v2 ), . . . , T (vn )} W is a basis for R(T ). Of course, to show this,
we need to show that this set both spans R(T ) and is linearly independent (recall Definition
2.10):
v = 1 v1 + 2 v2 + + n vn ,
Thus, every vector in R(T ) (i.e. every vector in W of the form T (v) for some v V )
can be written as a linear combination of the vectors in S 00 and so this set spans R(T ) (as
required).
To show that S 00 is linearly independent: As (T ) = 0, by Definition 2.16, the null
space of T is {0}, and as such, T (v) = 0 iff v = 0.15 Also, as the set {v1 , v2 , . . . , vn } is a
basis, we know that
v = 0 = 1 v1 + 2 v2 + + n vn iff 1 = 2 = = n = 0,
as these vectors are linearly independent. Consequently, using these two facts, we have
T (v) = 0 iff 1 = 2 = = n = 0,
in Equation (1), and so the vectors in S 00 are linearly independent (as required).
Thus, S 00 is a basis of R(T ), and as this set contains n vectors, (T ) = n. So, in this case, the
rank-nullity theorem holds since
(T ) + (T ) = 0 + n = n = dim(V ),
14
Case 2: Assume that (T ) = n. We further assume (as before) that dim(V ) = n and as such,
we can take the set of vectors S 0 = {v1 , v2 , . . . , vn } to be a basis for V . Now, as the null space
of T is a subspace of V (see Theorem 3.7) and dim(V ) = (T ) = n, we can see that V = N (T )
(see Problem Sheet 1, Question 12). Thus, S 0 will be a basis for N (T ) too, and so we have
Now, to prove the rank-nullity theorem in this case we need to find (T ), and to do this, we
need to find the range of T . But, R(T ) consists of the set of all vectors in W that are the image
of some vector v V , so taking such a general vector in V , say
v = 1 v1 + 2 v2 + + n vn ,
we can see that, by Theorem 3.2, R(T ) is the set of all vectors of the form
However, by Equation (2), this entails that T (v) = 0 for all such v V and so we have
R(T ) = {0}. Hence, by Definition 2.16, (T ) = 0. So, in this case, the rank-nullity theorem
holds since
(T ) + (T ) = n + 0 = n = dim(V ),
as we assumed that V was an n-dimensional space in the proof.
Finally, we note that the proof given in the handout does not apply in these cases because:
In Case 1, we assume that (T ) = 0 and as such the null space of T has no basis,16 whereas
the proof in the handout assumes that we do have a basis for the null space of T .
In Case 2, we find that (T ) = n and as such a basis for the null space of T is already a basis
for V , whereas the proof in the handout assumes that we can construct a basis for V by adding
[linearly independent] vectors to a basis for the null space.17
12. We are asked to prove Theorems 3.12 and 3.13 from the hand-out. There is really no need for
me to do this here as the proofs are essentially the arguments given at the beginning of Sections 3.3.3
and 3.3.4 respectively. If you are interested, you can try and reconstruct the proofs yourself.
16
To see why this is the case, see the discussion that immediately follows Definition 2.16.
17
And, in turn, these new vectors allow us to go on and generate a basis for R(T ), although in Case 2, the range of
T has no basis since R(T ) = {0}.
15