• Aucun résultat trouvé

1 Vector Pad´ e approximants of

N/A
N/A
Protected

Academic year: 2022

Partager "1 Vector Pad´ e approximants of "

Copied!
10
0
0

Texte intégral

(1)

Vector Pad´e approximants and the Lanczos method for solving a system of linear equations

Herv´e Le Ferrand

Laboratoire d’Analyse Appliqu´ee etd’Optimisation D´epartement de Math´ematiques

BP 138, 21004 Dijon cedex FRANCE

May 1994

Abstract

In the present paper we solve a system of linear equations by using the vectorPad´e approximants of a matrix series. The connection with the Lanczos methodand the oblique projections on Krylov subspaces is made.

Key words : Lanczos method, vector orthogonal polynomials, vector Pad´e approximants, minimal polynomial, biorthogonality.

Introduction

The vector Pad´e approximants considered in this paper are those defined by Van Iseghem in [14]. Their definition and properties will be summarized in the section 1. Let us consider the function𝐹(𝑧) =𝑛≥0𝑧𝑛𝐴𝑛𝑋0 from𝐶to𝐶𝑑where𝐴is a square𝑑×𝑑matrix and𝑋0 ∈𝐶𝑑. For 𝑧 small enoughwe have𝐹(𝑧) = (𝐼 −𝑧𝐴)−1 𝑋0. This function is rational and thus, some of the vector Pad´e approximants of the series will be identical to the series itself [9].On the other hand, the vector (𝐼−𝑧𝐴)−1 𝑋0 is the solution of the linear system𝑋 =𝑧𝐴𝑋+𝑋0. The vector Pad´e approximants of the series can be computed recursively, that gives a way to obtain the solution of the previous system. This method is related to a generalization to the vector case of the Shanks transform [13].Recently, Brezinski [4] generalized the method of moments of Vorobyev [15] and gave some generalizations of the Lanczos method for solving system of linear equations.

The aim of this paper is to show the connection between the vector Pad´e approximants and the projections on Krylov subspaces. The vector orthogonal polynomials associated to the function 𝑧 7→(𝐼−𝑧𝐴)−1 𝑋0 can be seen as the minimal polynomials of nonderogatory linear mappings into Krylov subspaces.This paper is organized as follows. In section 1 we deal with the linear system 𝑋 = 𝐴𝑋 +𝑋0. We show how to solve it by using vector Pad´e approximants of the resolvent of A applied to the vector 𝑋0. In section 2 we prove that a Lanczos type method is obtained by using vector orthogonal polynomials. Finally, the section 3 contains remarks about the different assumptions made before. Particularly, we make clearly the link between the Lanczos type method and the oblique projections on Krylov subspaces.

(2)

1 Vector Pad´ e approximants of 𝑧 7→ (𝐼 − 𝑧𝐴)

−1

𝑋

0

.

The notations are as follows.If 𝑝 𝑎𝑛𝑑 𝑞are two integers, [𝑝:𝑞] is the integer part of [𝑝𝑞].[𝑝, 𝑞] = 𝑃𝑄 is the vector Pad´e approximant of degree (𝑝, 𝑞) of the power series𝐹, with coefficients in 𝐶𝑑, i.e we have

𝑃 = (𝑃1, . . . , 𝑃𝑑),

𝑎𝑙𝑙 𝑃𝑖 𝑎𝑟𝑒 𝑝𝑜𝑙𝑦𝑛𝑜𝑚𝑖𝑎𝑙𝑠 𝑜𝑓 𝑑𝑒𝑔𝑟𝑒𝑒 𝑝, 𝑄 𝑖𝑠 𝑎 𝑝𝑜𝑙𝑦𝑛𝑜𝑚𝑖𝑎𝑙 𝑜𝑓 𝑑𝑒𝑔𝑟𝑒𝑒 𝑞, 𝑎𝑛𝑑

𝐹(𝑧)−[𝑝, 𝑞]𝐹(𝑧) =𝑂(𝑧𝑝 + [𝑞 :𝑑] + 1). (1) The polynomial 𝑃˜𝑟𝑠 (i.e 𝑃˜𝑟𝑠 = 𝑥𝑟𝑃𝑟𝑠(1𝑥) is the denominator of the vector Pad´e approximant [𝑟 +𝑠− 1/𝑟] (we assume of course, that the generalized Hankel’s determinant [12] doesn’t vanish).Let us now consider a linear operator 𝐴 of 𝐶𝑑. If necessary, we will identify A to its matrix, still denoted by 𝐴, in the canonical basis of 𝐶𝑑.

1.1

Let F be the function𝑧7→(𝐼−𝑧𝐴)−1 𝑋0. We assume that 𝑡ℎ𝑒 𝐾𝑟𝑦𝑙𝑜𝑣 𝑠𝑢𝑏𝑠𝑝𝑎𝑐𝑒 𝑠𝑝𝑎𝑛(𝑋0, 𝐴𝑋0, . . . , 𝐴𝑑1𝑋0

) 𝑖𝑠 𝑒𝑞𝑢𝑎𝑙 𝑡𝑜 𝐶𝑑. It was proved in [9] that 𝑃𝑑0 exists and is unique, and

𝑃𝑑0= Π𝐴=𝑃𝐴= Π𝐴,𝑋0

where Π𝐴 is the minimal polynomial of 𝐴, 𝑃𝐴 is the characteristic polynomial of A, Π𝐴,𝑋0 is the minimal polynomial of 𝐴 for the vector 𝑋0, and 𝑃𝑑0 is the vector orthogonal polynomial associated with the vector Pad´e approximant [𝑑−1/𝑑]𝐹. Moreover, since 𝑠= 0, we have even if A is singular

(𝐼−𝑧𝐴)−1𝑋0 = [𝑑−1/𝑑]𝐹(𝑧). (2)

Indeed we have

[𝑑−1/𝑑]𝐹(𝑧) =

𝑑

𝑗=1𝛼𝑗𝑧𝑑−𝑗𝑗−1𝑘=0Γ𝑘𝑧𝑘

𝛼0𝑧𝑑+⋅ ⋅ ⋅+𝛼𝑑 (3) (Γ𝑘=𝐴𝑘𝑋0 𝑎𝑛𝑑 𝑡ℎ𝑒 𝛼𝑖𝑠 𝑎𝑟𝑒 𝑡ℎ𝑒 𝑐𝑜𝑒𝑓 𝑓 𝑖𝑐𝑖𝑒𝑛𝑡𝑠 𝑜𝑓 𝑃𝑑0).

But

𝑑

𝑗=1

𝛼𝑗𝑧𝑑−𝑗(

𝑘≥𝑗

Γ𝑘𝑧𝑘) =

𝑑

𝑗=1

𝛼𝑗𝑧𝑑−𝑗(

𝑘≥0

Γ𝑘+𝑗𝑧𝑘+𝑗) (4)

= 𝑧𝑑(

𝑑

𝑗=1

𝛼𝑗𝐴𝑗)(

𝑘≥0

Γ𝑘𝑧𝑘) (5)

= 𝑧𝑑(−𝛼0)(𝐼−𝑧𝐴)−1𝑋0 (6)

(3)

and thus

(𝐼−𝑧𝐴)−1𝑋0 =

𝑑

𝑗=0𝛼𝑗𝑧𝑑−𝑗(𝑘≥0Γ𝑘𝑧𝑘)

𝑑

𝑗=0𝛼𝑗𝑧𝑑−𝑗 (7)

=

𝑗=1𝛼𝑗𝑧𝑑−𝑗(𝑘≥0Γ𝑘𝑧𝑘) − 𝛼0𝑧𝑑(𝐼−𝑧𝐴)−1𝑋0

𝑑

𝑗=0𝛼𝑗𝑧𝑑−𝑗 (8)

=

𝑑

𝑗=1𝛼𝑗𝑧𝑑−𝑗(𝑗−1𝑘=0Γ𝑘𝑧𝑘)

𝑑

𝑗=0𝛼𝑗𝑧𝑑−𝑗 (9)

= [𝑑−1/𝑑]𝐹(𝑧). (10)

Remark 1.1 If I-A is nonsingular, we get

(𝐼−𝐴)−1𝑋0 = [𝑑−1/𝑑]𝐹(1) =𝑄𝑑−1(𝐴)𝑋0 (11) where 𝑄𝑑−1 is the polynomial of degree 𝑑−1 defined by

1− 𝑃𝑑0(𝑡)

𝑃𝑑0(1) = (1−𝑡)𝑄𝑑−1(𝑡) (clearly 𝑑𝑗=1𝛼𝑗(𝑗−1𝑘=0𝐴𝑘) =𝑑−1𝑗=0(𝑑𝑘=𝑗+1𝛼𝑗)𝐴𝑗 and𝑄𝑑−1(𝑡) =

𝑑−1 𝑗=1(𝑑

𝑘=𝑗+1𝛼𝑘𝑡𝑗 𝛼0+𝛼1+⋅⋅⋅+𝛼𝑑 ).

1.2

More generally, let 𝑟 be an integer smaller or equal to 𝑑, the dimensionof the space. Let us assume that

𝑠𝑝𝑎𝑛(𝑋0, 𝐴𝑋0, . . . , 𝐴𝑟−1𝑋0) =𝑠𝑝𝑎𝑛(𝑒1, . . . , 𝑒𝑟) =𝐸𝑟, (𝑒𝑖)𝑑𝑖=1 𝑏𝑒𝑖𝑛𝑔 𝑡ℎ𝑒 𝑐𝑎𝑛𝑜𝑛𝑖𝑐𝑎𝑙 𝑏𝑎𝑠𝑖𝑠 𝑜𝑓 𝐶𝑑. According to [9], the vector Pad´e approximant [𝑟−1/𝑟]𝐹(𝑧) of𝐹(𝑧) = (𝐼−𝑧𝐴)−1𝑋0 exists and is unique. Its generating polynomial, i.e 𝑃𝑟0, is the characteristic polynomial of the operator 𝐴𝑟=𝛿𝑟𝐴 restricted to the subspace 𝐸𝑟 (𝛿𝑟 denotes the orthogonal projection operator on 𝐸𝑟).

The operator 𝐴𝑟 is the operator which appears in the moments method (see [15]). Moreover, we have

[𝑟−1/𝑟]𝐹(𝑧) = (𝐼𝑟−𝑧𝐴𝑟)−1𝑋0 (12) (𝐼𝑟 is the identity operator of 𝐸𝑟). Indeed, after noticing that

𝐴𝑘𝑟 =𝐴𝑘𝑓 𝑜𝑟 𝑘= 0, . . . , 𝑟−1,

we work on 𝐸𝑟 and apply the result of the section 1.1 which is true for every finite space, and for every nonderogatory operator.

If𝐼−𝐴is nonsingular, when we are allowed to choose z=1 and the remark 1.1 is still valid with, of course, 𝑟 replacing𝑑.

(4)

Remark 1.2 If, for example, we assume that

𝑠𝑝𝑎𝑛(𝑋0, 𝐴𝑋0, . . . , 𝐴𝑟−1𝑋0) = 𝑠𝑝𝑎𝑛(𝑒1, . . . , 𝑒𝑟) 𝑓 𝑜𝑟 𝑎𝑙𝑙 𝑟, 𝑟= 1 𝑡𝑜 𝑑, the canonical basis should mean the Gram-Schmidt basis.

We have now to see for which condition𝑃𝑟0(1) doesn’t vanish. Let us give the following condition Lemma 1.1 Let 𝐵 be an operator of 𝐶𝑑 and 𝑋0 ∈𝐶𝑑.

Let us assume that

𝑠𝑝𝑎𝑛(𝑋0, 𝐵𝑋0, . . . , 𝐵𝑟−1𝑋0) =𝐸𝑟.

Then the operator 𝐵𝑟, i.e 𝛿𝑟𝐵 restricted to 𝐸𝑟, is nonsingular if and only if

(𝑋0, 𝐵𝑋0) . . . (𝑋0, 𝐵𝑟𝑋0)

... ...

(𝐵𝑟−1𝑋0, 𝐵𝑋0) . . . (𝐵𝑟−1𝑋0, 𝐵𝑟𝑋0)

∕= 0 (13)

Proof

Let 𝑋=𝑟−1𝑖=0𝑥𝑖𝐵𝑖𝑋0 ∈𝐸𝑟, then we have𝐵𝑋 =𝑟−1𝑖=0 𝑥𝑖𝐵𝑖+1𝑋0. Thus 𝛿𝑟(𝐵𝑋) = 0 if and only if

𝑓 𝑜𝑟 𝑘= 0, . . . , 𝑟−1 (𝐵𝑘𝑋0, 𝐵𝑋) = 0, i.e

∀𝑘∈ {0, . . . , 𝑟−1}

𝑟−1

𝑖=0

𝑥𝑖(𝐵𝑘𝑋0, 𝐵𝑖+1𝑋0) = 0.

𝐵𝑟, operator de 𝐸𝑟, is then nonsingular if and only if the linear system

𝑟−1

𝑖=0

𝑥𝑖(𝐵𝑘, 𝐵𝑖+1), 𝑘= 0, . . . , 𝑟−1

is a nonsingular system and so we have the condition of the lemma. We need also Lemma 1.2 Let 𝐵 be an operator of 𝐶𝑑 and 𝑋0 ∈𝐶𝑑.

Let us assume that the vectors 𝑋0, 𝐵𝑋0, . . . , 𝐵𝑟−1𝑋0 are linearly independent. Then 𝑠𝑝𝑎𝑛(𝑋0, 𝐵𝑋0, . . . , 𝐵𝑋𝑟−1𝑋0) =𝑠𝑝𝑎𝑛(𝑋0,(𝐼−𝐵)𝑋0, . . . ,(𝐼−𝐵)𝑟−1𝑋0).

Proof

The proof is quite trivial by virtue of the binomial formula.Let us denote by 𝐴𝑟 the operator 𝛿𝑟𝐴restricted to 𝐸𝑟. By using the lemma 1.1 and the lemma 1.2, we obtain finally the

Theorem 1.1 Let 𝐴 be an operator of 𝐶𝑑 and 𝑋0 ∈𝐶𝑑. Let us asumme that 𝑠𝑝𝑎𝑛(𝑋0, 𝐴𝑋0, . . . , 𝐴𝑟−1𝑋0) =𝐸𝑟.

(5)

The generating polynomial 𝑃𝑟0 of [𝑟−1/𝑟]𝐹 is also the characteristic polynomial of 𝐴𝑟.If

(𝑋0,(𝐼−𝐴)𝑋0 . . . (𝑋0,(𝐼−𝐴)𝑟𝑋0)

... ...

((𝐼−𝐴)𝑟−1𝑋0,(𝐼−𝐴)𝑋0) . . . ((𝐼 −𝐴)𝑟−1𝑋0,(𝐼−𝐴)𝑟𝑋0)

∕= 0 (14) then 𝑃𝑟0(1)∕= 0 (i.e 𝐼𝑟−𝐴𝑟 is nonsingular) and if 𝑄𝑟−1 denotes the polynomial defined by

1− 𝑃𝑟0(𝑡)

𝑃𝑟0(1) = (1−𝑡)𝑄𝑟−1(𝑡), then

(𝐼𝑟−𝐴𝑟)−1𝑋0 = [𝑟−1/𝑟]𝐹(1) (15)

= 𝑄𝑟−1(𝐴𝑟)𝑋0 (16)

= 𝑄𝑟−1(𝐴)𝑋0. (17)

1.3

Let us give the following example 𝐴=

1 3 1 1 2 4 0 1 1

, 𝑋0 = (1,0,0)𝑇. We have

(𝐼−𝐴)−1𝑋0= (4,0,−1)𝑇, 𝐴𝑋0= (1,1,0)𝑇, 𝐴2𝑋0 = (4,3,1)𝑇, 𝐴3𝑋0 = (14,14,4)𝑇, 𝐴4𝑋0= (80,58,18)𝑇.

It is easy to see that the assumptions of the theorem 1.1 are satisfied. By using directly the determinantal expressions of the vector Pad´e approximants (see [14]), we obtain

[0/0] (𝑡) = (1,0,0)𝑇 (18)

[1/0] (𝑡) = (1 +𝑡, 𝑡,0)𝑇 (19)

[1/1] (𝑡) =

(3𝑡−1 4𝑡−1, −𝑡

4𝑡−1,0 )𝑇

(20) [1/2] (𝑡) =

( −2𝑡+ 1

−𝑡2−3𝑡+ 1, 𝑡

−𝑡2−3𝑡+ 1,0 )𝑇

(21) [2/1] (𝑡) =

(−𝑡2+ 5𝑡−2

7𝑡−2 ,𝑡2−2𝑡 7𝑡−2, −2𝑡2

7𝑡−2 )𝑇

(22) [2/2] (𝑡) =

(10𝑡2+ 5𝑡−2

14𝑡2−1 ,−3𝑡2−𝑡

14𝑡2−1, −𝑡2 14𝑡2−1

)𝑇

(23) [3/2] (𝑡) =

(−20𝑡3−12𝑡2−19𝑡+ 14

−14𝑡2−26𝑡+ 14 , 6𝑡3−5𝑡2+ 14𝑡

−14𝑡2−26𝑡+ 14, 2𝑡3+ 14𝑡2

−14𝑡2−26𝑡+ 14 )𝑇

(24) [2/3] (𝑡) =

( 2𝑡2+ 3𝑡−1

−4𝑡3+ 2𝑡2+ 4𝑡−1, 𝑡2−𝑡

−4𝑡3+ 2𝑡2+ 4𝑡−1, −𝑡2

−4𝑡3+ 2𝑡2+ 4𝑡−1 )𝑇

(25)

(6)

and

[2/3](1) = (4,0,−1)𝑇.

Remark 1.3 The vector Pad´e approximants can be also computed by using the generalizedcross rule (see [11]). Numerical examples using this rule can be found in [13].

In the generalized cross rule, determinants of dimension d occur. Thus, it will be difficult to compute them when d increases. Other algorithms can be used to compute the vector Pad´e approximants (see [5]). Let us mention for example, the recursive projection algorithm (R.P.A).

The same problem of implementation occurs in the initialization of the vector Q.D algorithm ([12]) (in [9], an initialization using the R.P.A is given).

2 The Lanczos method.

2.1

Let us now consider the linear system 𝐴𝑋 = 𝑋0 (𝐴 is nonsingular). It can be written as (𝐼−𝐵)𝑋=𝑋0 with𝐵 =𝐼−𝐴.

First of all, let us give the following result

Lemma 2.1 Let Π𝐴 be the minimal polynomial of 𝐴 and Π𝐵 be the minimal polynomial of 𝐵.

𝐼𝑓 𝐵=𝐼−𝐴, 𝑡ℎ𝑒𝑛 Π𝐵(𝑥) = Π𝐴(1−𝑥). (26) This result remains true if we consider the characteritic polynomials or the minimal polynomials for a vector.

Proof

This result comes from the Jordan decomposition of the matrix 𝐴 and from the notion of elementary polynomials, after having noticed that the spectrum of 𝐵 is the set {1−𝜆/𝜆 𝑒𝑖𝑔𝑒𝑛𝑣𝑎𝑙𝑢𝑒 𝑜𝑓 𝐴} (see [7]).We have now the

Theorem 2.1 Let𝐴 be a linear operator of 𝐶𝑑and 𝑋0 ∈𝐶𝑑. Let us denote by 𝐴𝑟 the operator 𝛿𝑟𝐴 restricted to 𝐸𝑟=𝑠𝑝𝑎𝑛(𝑒1, . . . , 𝑒𝑟).

Let us assume that

𝑠𝑝𝑎𝑛(𝑋0, 𝐴𝑋0, . . . , 𝐴𝑟−1𝑋0) =𝐸𝑟

and

(𝑋0, 𝐴𝑋0) . . . (𝑋0, 𝐴𝑟𝑋0)

... ...

(𝐴𝑟−1𝑋0, 𝐴𝑋0) . . . (𝐴𝑟−1𝑋0, 𝐴𝑟𝑋0)

∕= 0.

Then 𝐴𝑟 is nonsingular and the solution of the system 𝛿𝑟𝐴𝑋 =𝑋0, 𝑋 ∈𝐸𝑟 is

𝑋=𝑉𝑟−1(𝐴)𝑋0, (27)

where 𝑉𝑟−1 is the polynomial defined by the relation 1−𝑃𝑟𝑥

𝑃𝑟0 =𝑥𝑉𝑟−1(𝑥) (28)

(7)

with 𝑃𝑟0 the generating polynomial of the vector Pad´e approximant [𝑟 −1/𝑟] of the function 𝑧7→(𝐼−𝑧𝐴)−1𝑋0.

Proof

The nonvanishing condition of the previous determinant implies that𝐴𝑟is nonsingular.Moreover, we have

𝑠𝑝𝑎𝑛(𝑋0, 𝐴𝑋0, . . . , 𝐴𝑟−1𝑋0) =𝑠𝑝𝑎𝑛(𝑋0, 𝐵𝑋0, . . . , 𝐵𝑟−1𝑋0) =𝐸𝑟

and

𝐵𝑟=𝐼𝑟−𝐴𝑟.

Let us denote by 𝑃𝑟0 the generating polynomial of [𝑟 −1/𝑟]𝐹 where 𝐹(𝑧) = (𝐼 −𝑧𝐴)−1𝑋0 and by 𝑃 the generating polynomial of [𝑟−1/𝑟]𝐺 where 𝐺(𝑧) = (𝐼−𝑧𝐵)−1𝑋0. Applying the theorem 1.1 to 𝐵, we obtain (𝐼𝑟−𝐵𝑟)−1𝑋0 =𝑄(𝐵)𝑋0 where 𝑄 is the polynomial defined by 1−𝑃𝑃(𝑥)(1) = (1−𝑥)𝑄(𝑥).

Since P is the characteristic polynomial (and also the minimal polynomial) of𝐵𝑟, then it follows from lemma 2.1 that 𝑃(𝑥) =𝑃𝑟0(1−𝑥).

We deduce that

1−𝑃𝑟0(1−𝑥)

𝑃𝑟0(0) = (1−𝑥)𝑄(𝑥), and thus

1−𝑃𝑟0(𝑥)

𝑃𝑟0(0) =𝑥𝑄(1−𝑥) =𝑥𝑉𝑟−1(𝑥).

In conclusion, the following result is obtained

𝐴−1𝑟 𝑋0 = (𝐼𝑟−𝐵𝑟)−1𝑋0 =𝑉𝑟−1(𝐼−𝐵)𝑋0 (29) that is

𝐴−1𝑟 𝑋0=𝑉𝑟−1(𝐴)𝑋0 (30)

2.2

Let us now study the connection between the vector orthogonal polynomials and the Lanczos method.We consider in 𝐶𝑑 the linear system

(★) 𝐴𝑥=𝑏 (𝐴 𝑖𝑠 𝑛𝑜𝑛𝑠𝑖𝑛𝑔𝑢𝑙𝑎𝑟).

Let 𝑥0 ∈𝐶𝑑, we set 𝑟0 =𝑏−𝐴𝑥0. Throughout this section we are assuming that 𝑓 𝑜𝑟 𝑘= 1, . . . , 𝑚, 𝑠𝑝𝑎𝑛(𝑟0, 𝐴𝑟0, . . . , 𝐴𝑘−1𝑟0) =𝐸𝑘=𝑠𝑝𝑎𝑛(𝑒1, . . . , 𝑒𝑘) (m being the degree of the minimal polynomial of 𝐴 for the vector𝑟0).

Remark 2.1 This assumption seems very strong. In the section 3 some remark about this assumption will be made. Besides, an important example of such a matrix is an Hessenberg’s matrix.

(8)

Putting𝑥=𝑥0+𝑧, the system (★) becomes𝐴𝑧=𝑟0. Then we solve the linear system𝛿𝑘𝐴𝑧=𝑟0

in 𝐸𝑘. To do this, let us apply the theorem 2.1.

The condition

(𝑟0, 𝐴𝑟0) . . . (𝑟0, 𝐴𝑘𝑟0)

... ...

(𝐴𝑘−1𝑟0, 𝐴𝑟0) . . . (𝐴𝑘−1𝑟0, 𝐴𝑘𝑟0)

∕= 0 (31)

must be satisfied. Thus the solution of the system (★) is

𝑧𝑘=𝑉𝑘−1(𝐴)𝑟0 (32)

where 𝑉𝑘−1 is the polynomial defined by 1−𝑃𝑘0(𝑥)

𝑃𝑘0(0) =𝑥𝑉𝑘−1(𝑥), (33)

𝑃𝑘0being the generating polynomial of the vector Pad´e [𝑘−1/𝑘] of the function𝑧7→(𝐼−𝑧𝐴)−1𝑟0. Then we set

{ 𝑥𝑘 = 𝑥0+𝑧𝑘

𝑟𝑘 = 𝑏−𝐴𝑥𝑘 (34)

So it follows

𝑟𝑘 = 𝑏−𝐴𝑥𝑘 (35)

= 𝑏−𝐴(𝑥0+𝑉𝑘−1(𝐴)𝑟0) (36)

= 𝑟0−𝐴𝑉𝑘−1(𝐴)𝑟0 (37)

= 𝑃𝑘0(𝐴)

𝑃𝑘0(0)𝑟0 (38)

According to the definition of the vector orthogonal polynomials, we get see that 𝑃𝑘0(𝐴)𝑟0 is orthogonal to 𝐸𝑘,i.e

𝑟𝑘 𝑖𝑠 𝑜𝑟𝑡ℎ𝑜𝑔𝑜𝑛𝑎𝑙 𝑡𝑜 𝐸𝑘. Moreover, since 𝑟𝑚 = 0, we obtain

𝐴𝑥𝑚=𝑏. (39)

Remark 2.2 The computation of the 𝑃𝑘0 can be made, for example, by using the recurrence relations proved by Van Iseghem in [12]. But these polynomials don’t allow an implementation as easy as in the standard Lanczos method (see [10]). Thus, the connection with the Lanczos method as described above, is the theoritical aspect of the method of for solving a system of linear equations, of the section 1.

Let us now make some remarks about the different assumptions made before.

(9)

3 Biorthogonal polynomials. Oblique projection.

3.1 Biorthogonal polynomials.

Let us consider the most general biorthogonal polynomials, (𝑃𝑘)𝑘≥0, with the respect to the independent linear functionals (𝐿𝑖)𝑖≥0 on𝐶[𝑋]. They satisfy

𝐿𝑖(𝑃𝑘) = 0 𝑓 𝑜𝑟 𝑖= 0, . . . , 𝑘−1. (40) The vector orthogonal polynomials of dimension n are a special case, when the funtionals (𝐿𝑖)𝑖≥0

satisfy

𝐿𝑖(𝑥𝑗+1) =𝐿𝑖+𝑛(𝑥𝑗). (41)

In the sections 1 and 2, we considered the vector polynomials of dimension 𝑛=𝑑where𝑑is the dimension of the space on which we work. We only considered the polynomials (𝑃𝑘)𝑘≤𝑑. Thus we did not make use of the relation (41).

Further generalizations of the Lanczos method using biorthogonal polynomials are given in [4].

3.2 The assumption 𝑠𝑝𝑎𝑛(𝑋0, 𝐴𝑋0, . . . , 𝐴𝑘−1𝑋0) = 𝑠𝑝𝑎𝑛(𝑒1, . . . , 𝑒𝑘).

We still denote by (𝑒𝑖)𝑑𝑖=1 the canonical basis of𝐶𝑑. The orthogonal polynomials𝑃𝑘0 are defined by

(𝑒𝑖, 𝑃𝑘0(𝐴)𝑋0) = 0 𝑓 𝑜𝑟 𝑖= 1, . . . , 𝑘.

Let us try to express these relations as projection on the subspace𝑠𝑝𝑎𝑛(𝑋0, 𝐴𝑋0, . . . , 𝐴𝑘−1𝑋0).

If we assume that

𝑠𝑝𝑎𝑛(𝑋0, 𝐴𝑋0, . . . , 𝐴𝑘−1𝑋0) =𝑠𝑝𝑎𝑛(𝑒1, . . . , 𝑒𝑘),

this projection is the orthogonal projection on 𝑠𝑝𝑎𝑛(𝑋0, 𝐴𝑋0, . . . , 𝐴𝑘−1𝑋0). Moreover we immediatly have the matrix of the linear transformation 𝐴𝑘 in the basis (𝑒𝑖)𝑘𝑖=1. This is the matrix formed by the first 𝑘 rows and columns of the matrix 𝐴.Let us now assume that 𝑠𝑝𝑎𝑛(𝑋0, 𝐴𝑋0, . . . , 𝐴𝑘−1𝑋0) = 𝐸𝑘 with 𝑑𝑖𝑚𝐸𝑘 = 𝑘 (𝐸𝑘 is not necessary equal to 𝑠𝑝𝑎𝑛(𝑒1, . . . , 𝑒𝑘)).

Can we write

𝐶𝑑=𝐸𝑘⊕(𝑠𝑝𝑎𝑛(𝑒1, . . . , 𝑒𝑘)) ? It is necessary and sufficient that

𝐸𝑘(𝑠𝑝𝑎𝑛(𝑒1, . . . , 𝑒𝑘))={0}, which is equivalent to

(𝑒1, 𝑋0) . . . (𝑒1, 𝐴𝑘−1𝑋0)

... ...

(𝑒𝑘, 𝑋0) . . . (𝑒𝑘, 𝐴𝑘−1𝑋0)

∕= 0. (42)

We recover the condition of existence and uniqueness of the polynomial 𝑃𝑘0.

On the other hand, since the vectors 𝑋0, 𝐴𝑋0, . . . , 𝐴𝑘−1𝑋0 are linearly independent, this con- dition implies that the functionals 𝑒𝑖, 𝑖 = 1, . . . , 𝑘, considered as linear functionals on 𝐸𝑘 are

(10)

independent (see [6]). Thus, we can take the oblique projection on𝐸𝑘along (𝑠𝑝𝑎𝑛(𝑒1, . . . , 𝑒𝑘)). We obtain a nonderogatory operator𝐴𝑘of𝐸𝑘whose characteristic polynomial is𝑃𝑘0.Conclusion There is a strong connection between the Lanczos method for solving a system of linear equa- tions and the the vector Pad´e approximants. Particularly, the Lanczos and the vector Pad´e approximants give the same iterates when the matrix A is Hessenberg.References

[1] C.Brezinski, Acc´el´eration de la Convergence en Analyse Num´erique, Springer Verlag, Berlin, 1977.

[2] C.Brezinski, Pad´e type Approximations and General Orthogonal Polynomials, ISNM vol 50, Birkh¨auser, Basel, 1980.

[3] C.Brezinski, Biorthogonality and its Applications to Numerical Analysis, Marcel Dekker, New York, 1992.

[4] C.Brezinski, The methods of Vorobyev and Lanczos, to appear.

[5] C.Brezinski and M.Redivo-Zaglia, Extrapolation Methods. Theory and Practice. North Holland, Amsterdam, 1991.

[6] P.J Davis, Interpolation and Approximation, Dover, New York, 1975.

[7] F.R Gantmacher, Th´eorie des matrices, Dunod, Paris, 1966.

[8] G.H Golub and C.F Van Loan, Matrix Computations, The Johns Hopkins University Press, Baltimore, 1989, 2𝑛𝑑 edition.

[9] H. Le Ferrand, Vector orthogonal polynomials and matrix series, J.Comp.Appl.Math. 45 (1993), 267-282.

[10] Y.Saad, The Lanczos biorthogonalization algorithm and other oblique projection methods for solving large unsymmetric systems, SIAM J.Numer.Anal., 19(1982)485-506.

[11] J.Van Iseghem, An extended cross rule for vector Pad´e approximants, Applied Num.Math.2(1986)143-155.

[12] J.Van Iseghem, Vector orthogonal relation. Vector Q.D algorithm, J.Comp.Appl.Math.19(1987)141- 150.

[13] J.Van Iseghem, Convergence of Vectorial Sequences. Applications, to appear.

[14] J.Van Iseghem, Approximants de Pad´e vectoriels, Thesis, University of Lille 1.

[15] Yu. V.Vorobyev, Method of Moments in Applied Mathematics, Gordon and Breach, New York, 1965.

Références

Documents relatifs

L’accès aux archives de la série « Publications du Département de mathématiques de Lyon » im- plique l’accord avec les conditions générales d’utilisation

A BSTRACT - In this paper, we employ a new method to prove a representation for- mula related to the generalized Baouendi-Grushin vector fields, and then the Ostrowski type

2. b) Derive the least squares solution of the unknown parameters and their variance-covariance matrix.. Given the angle measurements at a station along with their standard

Using the method of moments, we prove that any polynomial mo- ment of the solution of the homogeneous Boltzmann equation with hard potentials or hard spheres is bounded as soon as

Gerold Alsmeyer, Bastien Mallein. A simple method to find all solutions to the functional equation of the smoothing transform.. In order to characterize all solutions within this

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des

Kuttner for his constructive criticism of an earlier version of the paper, which has resulted in the present version. Added in

In this section we prove the convergence of Algorithm I.1 using problem (3) for the working set selection (i.e. the algorithm used by SV M light ).. If Algorithm I.1 stops in