• Aucun résultat trouvé

MA1112: Linear Algebra II Dr. Vladimir Dotsenko (Vlad) Lecture 8

N/A
N/A
Protected

Academic year: 2022

Partager "MA1112: Linear Algebra II Dr. Vladimir Dotsenko (Vlad) Lecture 8"

Copied!
3
0
0

Texte intégral

(1)

MA1112: Linear Algebra II

Dr. Vladimir Dotsenko (Vlad) Lecture 8

Some examples for the case ϕ

k

= 0

Yesterday, we proved that for every linear transformation ϕ:V→V withϕk=0 for somek, it is possible to choose a basis

e(1)1 , e(1)2 , e(1)3 , . . . , e(1)n

1, e(2)1 , e(2)2 , . . . , e(2)n

2, . . .

e(l)1 , . . . , e(l)nl

ofV such that for each “thread”

e(p)1 , e(p)2 , . . . , e(p)np we have

ϕ(e(p)1 ) =e(p)2 , ϕ(e(p)2 ) =e(p)3 , . . . , ϕ(e(p)np) =0.

Today we shall discuss several examples of computing such bases of threads for a linear transformation.

Example 1. Let us consider the case ofV=R3, whereϕis multiplication by the matrixA=

−3 1 −1

−12 4 −4

−3 1 −1

.

We have A2 = 0, so ϕ2 = 0, falling into the class we considered. Note that rk(ϕ) = rk(A) = 1, so nullϕ=2.

We consider the sequence of subspaces V =Kerϕ2 ⊃Kerϕ⊃{0}. The first one relative to the second one is one-dimensional (since nullϕ2−nullϕ=3−2=1).

Note that the kernel of ϕ has a basis consisting of the vectors

 1/3

1 0

 and

−1/3 0 1

(corresponding to the valuess=1, t=0ands=0, t=1 of the free variables respectively). The reduced column echelon form of the corresponding matrixA=

1/3 −1/3

1 0

0 1

is the matrixR=

 1 0 0 1

−3 1

.

Since the basis vectors of Kerϕ have pivots in the first and the second row, it is easy to see that the vectorf=

 0 0 1

can be taken as a basis ofV relative to Kerϕ.

This vector gives rise to vector ϕ(f) =

−1

−4

−1

. It remains to find a basis of Kerϕ relative to the span

1

(2)

of ϕ(f). Column reduction of the basis vectors of Ker(ϕ) by ϕ(f) leaves us with the vector g =

 0 1 1

.

Overall,f, ϕ(f), gform a basis ofV. The matrix ofϕ relative to this basis is

0 0 0 1 0 0 0 0 0

.

Example 2. Now letV=R3, whereϕ is multiplication by the matrixA=

21 −7 8

60 −20 23

−3 1 −1

.

In this case,ϕ2is multiplication by the matrix

−3 1 −1

−9 3 −3

0 0 0

,ϕ3=0, rkϕ=2, rkϕ2=1, rkϕk=0

fork>3, null(ϕ) =1, null(ϕ2) =2, null(ϕk) =3 fork>3.

We consider the sequence of subspacesV=Kerϕ3⊃Kerϕ2⊃Kerϕ⊃{0}. The first one relative to the second one is one-dimensional (nullϕ3−nullϕ2=1).

The vector space Ker(ϕ2)has a basis consisting of the vectors

 1/3

1 0

and

−1/3 0 1

 (corresponding to the valuess=1, t=0ands=0, t=1 of the free variables respectively). The reduced column echelon form of the corresponding matrixA=

1/3 −1/3

1 0

0 1

is the matrixR=

 1 0 0 1

−3 1

.

Since the basis vectors of Kerϕ2 have pivots in the first and the second row, the vector f =

 0 0 1

can be taken as a basis of V relative to Kerϕ2. This vector gives rise to the thread f, ϕ(f) =

 8 23

−1

,

ϕ2(f) =

−1

−3 0

. Since our space is 3-dimensional, this thread forms a basis. The matrix of ϕ relative to

this basis is

0 0 0 1 0 0 0 1 0

.

Example 3. LetV=R4, where ϕis multiplication by the matrixA=

1 0 0 1

0 −1 1 0

0 −1 1 0

−1 0 0 −1

 .

In this case,ϕ2=0, rk(ϕ) =2, rk(ϕk) =0 fork>2, null(ϕ) =2, null(ϕk) =4 fork>2.

We consider the sequence of subspacesV=Ker(ϕ2)⊃Ker(ϕ)⊃{0}. The first one relative to the second one is two-dimensional (nullϕ2−nullϕ=2).

The vector space Ker(ϕ) has a basis consisting of the vectors

−1 0 0 1

 and

 0 1 1 0

(corresponding to the

valuess=1, t=0ands=0, t=1of the free variables respectively). The reduced column echelon form has

pivots in row one and row two, so the vectorsf1=

 0 0 1 0

andf2=

 0 0 0 1

can be taken as a basis ofV relative

2

(3)

to Ker(ϕ).These vectors give rise to threadsf1, ϕ(f1) =

 0 1 1 0

andf2, ϕ(f2) =

 1 0 0

−1

. These two threads together contain four vectors, so since our space is4-dimensional, we have a basis. The matrix ofϕrelative to this basis is

0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0

 .

3

Références

Documents relatifs

The key point of linear algebra is exactly that: an informed choice of a coordinate system can simplify things quite

We shall temporarily take the out- look which views symmetric matrices in the spirit of one of the proofs from last week, and looks at the associated bilinear forms (and

Today we shall prove various theorems on quadratic forms stated without proof last week..

Intuitively, orthogonal matrices with det(A) = 1 are transformations that can distinguish between left and right, or clockwise and counterclockwise (like rotations), and

We proved that every normal linear transformation admits an orthonormal basis of eigenvectors, so the first part follows.. This theorem can be used to deduce the theorem on

Given 101 coins of various shapes and denominations, one knows that if you remove any one coin, the remaining 100 coins can be divided into two groups of 50 of equal total weight..

, f n the “new basis”, then this system tells us that the product of the transition matrix with the columns of new coordinates of a vector is equal to the column of old

For a matrix A, the dimension of the solution space to the system of equations Ax = 0 is equal to the number of free unknowns, that is the number of columns of the reduced row