• Aucun résultat trouvé

On a fast Arnoldi method for BML matrices

N/A
N/A
Protected

Academic year: 2021

Partager "On a fast Arnoldi method for BML matrices"

Copied!
28
0
0

Texte intégral

(1)

HAL Id: hal-01451722

https://hal.archives-ouvertes.fr/hal-01451722

Preprint submitted on 1 Feb 2017

HAL is a multi-disciplinary open access archive for the deposit and dissemination of sci- entific research documents, whether they are pub- lished or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.

On a fast Arnoldi method for BML matrices

Bernhard Beckermann, Clara Mertens, Raf Vandebril

To cite this version:

Bernhard Beckermann, Clara Mertens, Raf Vandebril. On a fast Arnoldi method for BML matrices.

2017. �hal-01451722�

(2)

On a fast Arnoldi method for BM L-matrices

Bernhard Beckermann, Clara Mertens, and Raf Vandebril

Abstract

Matrices whose adjoint is a low rank perturbation of a rational function of the matrix naturally arise when trying to extend the well known Faber-Manteuffel theorem [7, 8], which provides necessary and sufficient conditions for the existence of a short Arnoldi recurrence.

We show that an orthonormal Krylov basis for this class of matrices can be generated by a short recurrence relation based on GMRES residual vectors. These residual vectors are computed by means of an updating formula. Furthermore, the underlying Hessenberg matrix has an accompanying low rank structure, which we will investigate closely.

1 Introduction

In this article we will discuss a new variant of the Arnoldi method applied to a class of sparse matrices A ∈ C

n×n

which allows to compute the first k Arnoldi vectors in complexity O(kn).

We will refer to this class of matrices as BML-matrices, following the fundamental work of Barth & Manteuffel [2] and Liesen [13] on matrices A whose adjoint is a low rank perturbation of a rational function of A. More specifically, we assume that

A

= p(A)q(A)

−1

+ F G

, (1.1)

with p, q polynomials of degree m

1

and m

2

, respectively, and

F = [f

1

, f

2

, . . . , f

m3

], G = [g

1

, g

2

, . . . , g

m3

] ∈ C

n×m3

matrices of full column rank. Moreover, it is assumed that the roots of q are simple. By taking p(z)/q(z) ∈ {z, 1/z}, we see that Hermitian matrices and unitary matrices are BM L- matrices, and the same is true for low rank perturbations of such matrices. Furthermore, if F G

= 0 in (1.1), the matrix A is normal [13]. In what follows we suppose that m

j

n, since these quantities (as well as the sparsity pattern of A) are hidden in the constant of the above-claimed complexity result.

After k steps of the Arnoldi process with initial vector b one obtains the expression AV

k

= V

k

H

k

+ h

k+1,k

v

k+1

e

k

, (1.2) with V

k

= [v

1

, v

2

, . . . , v

k

], v

1

= b/kbk, v

k+1

∈ C

n

and h

k+1,k

≥ 0 satisfying V

k

V

k

= I

k

and V

k

v

k+1

= 0. The matrix H

k

is upper Hessenberg.

A fast variant of the Arnoldi process will exploit additional structure of the upper Hes- senberg matrix H

k

. For example, to compute the successive vectors v

k

for Hermitian A we get a tridiagonal H

k

and the Arnoldi process reduces to the Lanczos method. For matrices A satisfying (1.1), H

k

turns out to be a rank structured matrix. In order to specify this statement, the following definition is introduced.

Definition 1.1. We say that a matrix B ∈ C

n×n

is (r, s)−upper-separable with s ≥ 0, if for all j = 1, 2, . . . , n − |r| it holds that (in Matlab notation)

rank B(1 : j − min(r, 0), j + max(r, 0) : n) ≤ s.

In other words, any submatrix of B including elements on and above the rth diagonal of B is

of rank at most s. The 0th diagonal corresponds to the main diagonal, while the rth diagonal

refers to the rth superdiagonal if r > 0 and to the −rth subdiagonal if r < 0.

(3)

Before proceeding, we give some comments on Definition 1.1 whose formulation is inspired by some related well-established definitions. For example, matrices with both B and B

being (0, 1)–upper-separable (and (1, 1)–upper-separable, respectively) are referred to as semisepa- rable matrices (and quasi separable, respectively) [18, §1.1, §9.3.1]. As a simple example, a tridiagonal matrix is quasi separable, and its inverse is known to be semiseparable. Matrices being (1 − s, s)–upper-separable with their adjoint being (1 − r, r)–upper-separable are usu- ally called (r, s)−semiseparable, while matrices being (1, s)–upper-separable with their adjoint being (1, r)–upper-separable are called (r, s)−quasi separable [18, §8.2.2 and §8.2.3].

The above statement on the rank structure of H

k

can now be made exact. Assume the Arnoldi process breaks down after N ≤ n iterations, i.e., v

N+1

= 0 in (1.2). We will refer to N as the ‘Arnoldi termination index’ in the rest of the article. In Corollary 3.2 it is shown that for a BM L-matrix A, the underlying Hessenberg matrix H

N

is (r, s)−upper-separable, where r and s are functions of m

j

. However, for a general matrix A, there is no reason for the underlying Hessenberg matrix to be upper-separable. The upper-separable structure of the underlying Hessenberg matrix gives rise to the design of a multiple recurrence relation [2, 14], signifying that each new Arnoldi vector can be written as

v

k+1

=

k

X

j=k−m3−m2

α

k,j

Av

j

+

k

X

j=k−m3−m1

β

k,j

v

j

. (1.3)

The smaller the quantities m

j

in (1.1), the shorter the recurrence relation becomes. In [5] the same recurrence relations are derived for the class of so called (H, m)-well-free matrices. We refer to Remark 3.7 for some more details.

In this article we investigate a different version of the recurrence relation (1.3) by rewriting it in terms of GMRES residual vectors, aiming to overcome some of the numerical problems which relation (1.3) entails, such as the possibility of a breakdown [14]. It will be shown how these GMRES residual vectors can be computed progressively by means of an updating formula. This partially extends the discussion on a progressive GMRES method for nearly Hermitian matrices as presented by Beckermann & Reichel [4].

The article is organized as follows. Section 2 describes the structure of the unitary factor Q in the QR-decomposition of a Hessenberg matrix in terms of orthogonal polynomials (the results in this section are valid for a general matrix A ∈ C

n×n

). It is well known that this unitary factor can be represented as a product of Givens rotations, see e.g., the isometric Arnoldi process introduced by Gragg [10] and its extension to the class of shifted unitary matrices by Jagels & Reichel [12]. We will describe these Givens rotations by means of orthogonal polynomials. Furthermore, links between orthogonal polynomials and the GMRES algorithm are discussed, leading to an updating formula to compute the GMRES residuals progressively. In section 3 the upper-separable structure of the Hessenberg matrix related to the Arnoldi process applied to a BM L-matrix is investigated. It is shown how this upper- separable structure can be generated by the GMRES residual vectors, allowing to construct a short recurrence relation and an accompanying algorithm. In section 4 we compare our findings with those presented in [2]. Section 5 discusses some computational reductions that can be made in case the matrix A is nearly unitary or nearly shifted unitary. Finally, section 6 discusses the numerical performance and stability of the algorithm.

Throughout this article we will make use of the following notation. Vectors are written in bold face lower case letters, e.g., x, y, and z. The vector e

k

denotes the kth column of an identity matrix of applicable order. The standard inner product is denoted as hx, yi = y

x, with ·

the Hermitian conjugate. We write k · k for the induced Euclidean norm as well as the subordinate spectral matrix norm. Matrices are denoted by upper case letters A = (a

ij

), and I

k

denotes the identity matrix of order k. We will frequently express formula (1.2) in the form

AV

k

= V

k+1

H

k

, with V

k+1

:= [V

k

, v

k+1

], where

I

k

:=

I

k

0

∈ C

(k+1)×k

, H

k

:=

H

k

h

k+1,k

e

k

= H

k+1

I

k

∈ C

(k+1)×k

,

and V

k

= V

k+1

I

k

, revealing the nested structure of the Arnoldi matrices V

k

and H

k

.

(4)

2 Towards a progressive GMRES method

Many Krylov space methods and in particular the Arnoldi method can be described in poly- nomial language which reveals some particular properties, and makes the link to the rich theory of orthogonal polynomials. For example, from (1.2) one deduces by recurrence on the degree that, for any polynomial p

p(A)v

1

= V

k

p(H

k

)e

1

, provided that deg p < k, (2.1) illustrating that Krylov spaces are intimately related to polynomials. See e.g., [16, §6.6.2]

for the case of Hermitian A. However, polynomial language can be used also for general matrices [3, §1.3]. In §2.1 we will see that the Arnoldi vectors correspond to a (finite) family of polynomials which are orthogonal with respect to

hp, qi

A,v1

= hp(A)v

1

, q(A)v

1

i = (q(A)v

1

)

p(A)v

1

, (2.2) a scalar product on the set of polynomials of degree < N , with N the termination index of the Arnoldi method. The scalar product (2.2) induces a norm k · k

A,v1

. This implies in particular the well known fact [16, Proposition 6.7] that Arnoldi vectors are normalized FOM residuals. In §2.2 we will use polynomial language to give an explicit expression for the Q- factor in a QR-decomposition of an upper Hessenberg matrix. Such a formula for a unitary upper Hessenberg matrix is not known in literature, though of course there is a close link with Gragg’s explicit formula in terms of Givens rotations [10, 16, §6.5.3]. This formula will enable us to deduce in Corollary 2.4 a decay property of the entries of Q far from the main diagonal, and reveals immediately a rank structure for Q. In §2.3 we recall that GMRES residuals can be expressed in terms of orthogonal polynomials: we are faced with a well-studied extremal problem for general orthogonal polynomials. This allows us in §2.4 to establish a well known link between (normalized) FOM and GMRES residuals [16, §6.5.5], allowing for a recursive computation of (normalized) GMRES residuals. Such a progressive GMRES implementation has been discussed before [16, §6.5.3] and [4]. The implementation presented in this article is inspired by the work of Beckermann & Reichel [4]. Both implementations make use of the decomposition of Q into a product of Givens rotations, but differ in how to find the angles of these rotations. We will consider in §2.4 not only FOM and GMRES for systems of linear equations Ax = b but more generally for shifted systems (A − δI)x = b for some parameters δ ∈ C . All findings of this section hold for general matrices A.

2.1 Orthogonal polynomials linked to the Arnoldi process

Given an N × N upper Hessenberg matrix H

N

with positive real entries h

k+1,k

on the sub- diagonal. We define polynomials q

0

, q

1

, ..., q

N−1

recursively through the formula

q

0

(z) = 1,

q

k

(z )h

k+1,k

= zq

k−1

(z) −

k

X

j=1

q

j−1

(z)h

j,k

, if N − 1 ≥ k ≥ 1. (2.3) Direct computation yields the following well known link with the characteristic polynomials of the principal submatrices:

q

k

(z) = 1 Q

k

i=1

h

i+1,i

det(zI

k

− H

k

), (2.4)

showing that q

k

is of degree k, with positive leading coefficient. In what follows we will write (2.3) in the form

(q

0

(z), . . . , q

k

(z))H

k

= z(q

0

(z), . . . , q

k−1

(z)), (2.5) where H

k

is the k × k principal minor of H

N

and N − 1 ≥ k ≥ 1. One deduces from (2.3) by recurrence on k that q

k

(H

N

)e

1

= e

k+1

. Assume we apply the Arnoldi process to a matrix A ∈ C

n×n

, that N ≤ n is the Arnoldi termination index and H

N

the underlying Hessenberg matrix. Using (2.1), this implies

v

k

= q

k−1

(A)v

1

, (2.6)

(5)

and thus indeed the q

j

is the jth orthonormal polynomial with respect to the scalar product (2.2). Also, from (2.4) we see that the kth FOM iterate for the shifted system (A − δI)x = b exists if and only if q

k

(δ) 6= 0, and that in this case the kth FOM residual is given by q

k

(A)b/q

k

(δ), compare with [16, Proposition 6.7].

2.2 The QR-factorization of a Hessenberg matrix

We will derive an explicit formula for the unitary factor in the QR-decomposition of the upper Hessenberg matrix H

k

in terms of the orthonormal polynomials q

0

, ..., q

k

. To our knowledge, such a result is new. It could also be potentially useful for studying the convergence of the QR-method with shifts.

Let Q

k+1

(δ) be the unitary factor in the QR-decomposition of H

k

− δI

k

, i.e., Q

k+1

(δ)

(H

k

− δI

k

) = R

k

(δ) :=

R

k

(δ) 0

∈ C

(k+1)×k

, (2.7)

R

k

(δ) an upper triangular matrix with positive real entries on its main diagonal. It is well known (see, e.g., [16, Subsection 6.5.3]) that Q

k+1

(δ) can be obtained as a product of Givens rotations, which are applied to H

k

− δI

k

to annihilate the first subdiagonal. We follow [4], imposing the matrices Q

k+1

(δ) to have determinant 1.

Definition 2.1. Let Q

1

= [1], and define for k ≥ 1,

Q

k+1

(δ)

= Ω

k+1

(δ)

Q

k

(δ)

0

0 1

, Ω

k+1

(δ) =

I

k−1

0 0

0 c

k

(δ) s

k

(δ) 0 −s

k

(δ) c

k

(δ)

 ,

with s

k

(δ) ≥ 0 and s

k

(δ)

2

+ |c

k

(δ)|

2

= 1, such that (2.7) holds.

Proposition 2.2. Let δ ∈ C , and

σ

k

(z) :=

v u u t

k

X

j=0

|q

j

(z)|

2

.

Then the unitary factor Q

k+1

(δ)

is given by

σq0(δ)q1(δ)

1(δ)σ0(δ)

σ0(δ)

σ1(δ)

0 · · · · · · 0

σq0(δ)q2(δ)

2(δ)σ1(δ)

σq1(δ)q2(δ)

2(δ)σ1(δ)

σ1(δ) σ2(δ)

0 . .

. . . . . . . . . . . . . . . .

. .

. . . . . . . . . . 0

σq0(δ)qk(δ)

k(δ)σk−1(δ)

σq1(δ)qk(δ)

k(δ)σk−1(δ)

· · · · · · −

σqk−1(δ)qk(δ)

k(δ)σk−1(δ)

σk−1(δ) σk(δ)

(−1)

k qσ0(δ)

k(δ)

(−1)

k qσ1(δ)

k(δ)

· · · · · · (−1)

k qk−1σ (δ)

k(δ)

(−1)

k qσk(δ)

k(δ)

 .

Moreover,

s

k

(δ) = σ

k−1

(δ)

σ

k

(δ) and c

k

(δ) = (−1)

k

q

k

(δ)

σ

k

(δ) , for k ≥ 1. (2.8)

Proof. We leave it to the reader to check that the candidate for Q

k+1

(δ)

indeed has orthonor-

mal rows. It remains to check the subdiagonal and diagonal entries of Q

k+1

(δ)

(H

k

− δI

k

).

(6)

For k ≥ j > ` we find that

e

j

Q

k+1

(δ)

(H

k

− δI

k

)e

`

= − q

j

(δ)

σ

j

(δ)σ

j−1

(δ) q

0

(δ), . . . , q

j−1

(δ), 0, . . . , 0

| {z }

k+1−j

)(H

k

− δI

k

)e

`

+

 0, . . . , 0

| {z }

j

, σ

j−1

(δ)

σ

j

(δ) , 0, . . . , 0

| {z }

k−j

 (H

k

− δI

k

)e

`

= − q

j

(δ)

σ

j

(δ)σ

j−1

(δ) (q

0

(δ), . . . , q

`

(δ)) (H

`

− δI

`

)e

`

= 0,

where the second equality is because of the fact that only the first ` + 1 ≤ j entries of (H

k

− δI

k

)e

`

are nonzero and the third equality because of (2.5). Similarly, for k ≥ j = `,

e

`

Q

k+1

(δ)

(H

k

− δI

k

)e

`

= − q

`

(δ)

σ

`

(δ)σ

`−1

(δ) (q

0

(δ), . . . , q

`−1

(δ), 0, . . . , 0

| {z }

k+1−`

)(H

k

− δI

k

)e

`

+

 0, . . . , 0

| {z }

`

, σ

`−1

(δ)

σ

`

(δ) , 0, . . . , 0

| {z }

k−`

 (H

k

− δI

k

)e

`

= − q

`

(δ)

σ

`

(δ)σ

`−1

(δ) (q

0

(δ), . . . , q

`−1

(δ))(H

`

− δI

`

)e

`

+ σ

`−1

(δ) σ

`

(δ) h

`+1,`

= − q

`

(δ)

σ

`

(δ)σ

`−1

(δ) (−q

`

(δ)h

`+1,`

) + σ

`−1

(δ) σ

`

(δ) h

`+1,`

= h

`+1,`

σ

`

(δ)σ

`−1

(δ) |q

`

(δ)|

2

+ σ

`−1

(δ)

2

= h

`+1,`

σ

`

(δ) σ

`−1

(δ) > 0.

Finally, for k + 1 = j ≥ `,

e

k+1

Q

k+1

(δ)

(H

k

− δI

k

)e

`

= (−1)

k

σ

k

(δ) (q

0

(δ), q

1

(δ), . . . , q

k

(δ)) (H

k

− δI

k

) e

`

= 0, according to (2.5). To prove (2.8), observe that

Q

k+1

(δ)

= Ω

k+1

(δ)

Q

k

(δ)

0

0 1

,

if and only if by multiplying on the left with

c

k

(δ) s

k

(δ)

−s

k

(δ) c

k

(δ)

we transform

"

(−1)

k−1σq0(δ)

k−1(δ)

(−1)

k−1σq1(δ)

k−1(δ)

· · · (−1)

k−1qσk−1(δ)

k−1(δ)

0

0 0 · · · 0 1

#

into

σq0(δ)qk(δ)

k(δ)σk−1(δ)

σq1(δ)qk(δ)

k(δ)σk−1(δ)

· · · −

σqk−1(δ)qk(δ)

k(δ)σk−1(δ)

σk−1(δ) σk(δ)

(−1)

k qσ0(δ)

k(δ)

(−1)

k qσ1(δ)

k(δ)

· · · (−1)

k qk−1σ (δ)

k(δ)

(−1)

k qσk(δ)

k(δ)

 , the latter being true for c

k

(δ) and s

k

(δ) as in (2.8).

Notice that, according to the nested structure of the Hessenberg matrices, Q

k+1

(δ)

(H

k+1

− δI

k+1

) is also upper triangular, but its last diagonal entry given by

(−1)

k+1

h

k+2,k+1

q

k+1

(δ)/σ

k

(δ)

(7)

is not necessarily a positive real number. Hence, for obtaining the unitary factor in the unique QR-decomposition of H

k+1

− δI

k+1

we should rescale the last row of Q

k+1

(δ)

as given in Proposition 2.2 by a phase of modulus 1.

Let us consider the special case of δ = 0 and unitary A as a running example.

Example 2.3. Suppose that δ = 0 and A is unitary. Then AV

k

= V

k+1

H

k

and thus H

k

has orthonormal columns, showing that H

k

= Q

k+1

(0)I

k

and R

k

(0) = I

k

. From Proposition 2.2 we get explicit formulas for the entries of H

k

, in particular, for unitary A,

h

k+1,k

= σ

k−1

(0)

σ

k

(0) = s

k

(0), (−1)

k−1

σ

k−1

(0)h

1,k

= (−1)

k

q

k

(0)

σ

k

(0) = c

k

(0). (2.9) Furthermore, AV

k

= V

k+1

Q

k+1

(0)I

k

. Therefore,

Av

k

= V

k+1

Q

k

(0) 0

0 1

I

k−1

0 0 c

k

(0) 0 s

k

(0)

 e

k

= V

k+1

c

k

(0)Q

k

(0)e

k

s

k

(0)

= s

k

(0)v

k+1

+ (−1)

k−1

c

k

(0)˜ v

k

, with ˜ v

k

:=

σ 1

k−1(0)

P

k−1

j=0

q

j

(0)v

j+1

. Also, ˜ v

k+1

= s

k

(0)˜ v

k

+ (−1)

k

c

k

(0)v

k+1

. Hence, the orthonormal vectors v

k

can be constructed using two short recurrence relations:

˜ v

1

:= v

1

;

for k = 1, . . . , N − 1 do z := Av

k

;

γ

k

:= −˜ v

k

z;

σ

k

:= (1 − |γ

k

|

2

)

1/2

; v

k+1

:= σ

k−1

(z + γ

k

v ˜

k

);

˜

v

k+1

:= σ

k

˜ v

k

+ γ

k

v

k+1

; end

Note that γ

k

= (−1)

k

c

k

(0) and σ

k

= s

k

(0). The above double recurrence relation is known as the ‘Isometric Arnoldi algorithm’ designed by Gragg [10].

As a consequence of Proposition 2.2, according to Definition 1.1, we can derive some statements on the rank structure of the unitary Hessenberg matrix Q

k+1

(δ), and on a decay property of its entries.

Corollary 2.4. Q

k+1

(δ) is (0, 1)-upper-separable

1

. Moreover, for the submatrix Q e of Q

k+1

(δ) formed with the first m ≤ k + 1 rows and the last ` ≤ k − m + 2 columns we have that

k Qk e = σ

m−1

(δ)/σ

k−`+1

(δ).

Proof. The first statement follows by observing that

Q

k+1

(δ) =

 q

0

(δ)

. . . q

k

(δ)

e

1

Q

k+1

(δ) + L

k+1

(δ), (2.10)

with L

k+1

(δ) strictly lower triangular, i.e., with zero entries on the main diagonal. This is a direct consequence of Proposition 2.2. In particular, we deduce that Q e is of rank 1. From Proposition 2.2 it follows that

k Qk e

2

= σ

m−1

(δ)

2

σ

k

(δ)

2

+

k

X

j=k−`+2

|q

j

(δ)|

2

σ

j

(δ)

2

σ

j−1

(δ)

2

σ

m−1

(δ)

2

= σ

m−1

(δ)

2

σ

k

(δ)

2

+

k

X

j=k−`+2

σ

j

(δ)

2

− σ

j−1

(δ)

2

σ

j

(δ)

2

σ

j−1

(δ)

2

σ

m−1

(δ)

2

= σ

m−1

(δ)

2

 1 σ

k

(δ)

2

+

k

X

j=k−`+2

1

σ

j−1

(δ)

2

− 1 σ

j

(δ)

2

 ,

1This implies thatQk+1(δ) is quasi separable, but in general not semiseparable.

(8)

giving the claimed result.

We end this subsection by observing that Corollary 2.4 immediately implies a rank prop- erty as well as a decay of entries for resolvents of H

k+1

.

Corollary 2.5. Suppose that H

k+1

− δI

k+1

is invertible. Then (H

k+1

− δI

k+1

)

−∗

is (0, 1)- upper-separable.

Moreover, for the submatrix H e of (H

k+1

− δI

k+1

)

−∗

formed with the first m ≤ k + 1 rows and the last ` ≤ k − m + 2 columns we have that

k Hk ≤ k(H e

k+1

− δI

k+1

)

−1

k σ

m−1

(δ)/σ

k−`+1

(δ).

Proof. Let us write H

k+1

− δI

k+1

= Q

k+1

(δ)R with upper triangular and invertible R. Then (H

k+1

− δI

k+1

)

−∗

= Q

k+1

(δ)R

−∗

with R

−∗

lower triangular of norm k(H

k+1

− δI

k+1

)

−1

k. Replacing Q

k+1

(δ) by (2.10) yields (H

k+1

− δI

k+1

)

−∗

= (q

0

(δ), . . . , q

k

(δ))

e

1

(H

k+1

− δI

k+1

)

−∗

+ L e

k+1

(δ), (2.11) with L e

k+1

(δ) strictly lower triangular; proving the first statement. This implies that

H e = (q

0

(δ), ..., q

m−1

(δ))

e

1

H, e

and thus k H e k = σ

m−1

(δ) ke

1

Hk. Notice that e e

1

H e is obtained by multiplying the first row of Q

k+1

(δ) with the last ` columns of R

−∗

. As R

−∗

is lower triangular, e

1

H e = Q e R, e Q e a row vector formed with the last ` entries of the first row of Q

k+1

(δ) and R e the lower-right ` × ` minor of R

−∗

. Therefore, applying Corollary 2.4 to Q e yields

ke

1

H e k ≤ || Q|| || e R|| ≤ e σ

0

(δ)

σ

k−`+1

(δ) ||R

−∗

||,

which together with k Hk e = σ

m−1

(δ) ke

1

Hk e and kR

−∗

k = k(H

k+1

− δI

k+1

)

−1

k proves the second statement.

2.3 The GMRES residual and orthogonal polynomials

We will give some more details on the link between the GMRES residual vectors and orthog- onal polynomials. More specifically, we will write the GMRES residual vector as a linear combination of Arnoldi vectors. In the particular case of unitary A, an even nicer relation arises for the GMRES residual vectors.

In what follows we denote by r

k

(δ) the kth GMRES residual for the shifted system (A − δI

n

)x = b, with starting vector x

0

= 0, and denote by w

k

(δ) := r

k

(δ)/kr

k

(δ)k its normalized version.

Proposition 2.6. The kth GMRES residual for the shifted system (A − δI

n

)x = b with starting vector x

0

= 0, can be written as

r

k

(δ) = p

k

(A)r

0

(δ), where p

k

(z) = 1 σ

2k

(δ)

k

X

j=0

q

j

(δ)q

j

(z). (2.12)

For its normalized version we have

w

k

(δ) = r

k

(δ) kr

k

(δ)k = 1

σ

k

(δ)

k

X

j=0

q

j

(δ)v

j+1

. (2.13)

Proof. Since we choose as starting vector x

0

= 0, we find the initial GMRES residual r

0

(δ) = b − (A − δI

n

)0 = b = v

1

kbk. Then we have

r

k

(δ) = r

0

(δ) − (A − δI

n

)V

k

y, with y = argmin

y

kr

0

(δ) − (A − δI

n

)V

k

yk. (2.14)

(9)

Note that r

0

− (A − δI

n

)V

k

y can be written as p

k

(A)r

0

(δ) with p

k

a polynomial of degree at most k and p

k

(δ) = 1. Then r

k

(δ) = p

k

(A)r

0

, p

k

= p e

k

/ p e

k

(δ), with

p e

k

= argmin

pPk

kp(A)r

0

(δ)k

|p(δ)| = argmin

p∈Pk

kpk

A,v1

|p(δ)| , (2.15) where P

k

denotes the set of polynomials of degree at most k, and we use the norm induced by (2.2). Note that if p(z) = P

k

j=0

c

j

q

j

(z), then kpk

2A,v1

= P

k

j=0

|c

k

|

2

by orthonormality.

Therefore, the Cauchy-Schwarz inequality yields kpk

A,v1

|p(δ)| ≥ 1

σ

k

(δ) , (2.16)

where the minimum is attained for c

j

= q

j

(δ). Combining (2.16) and (2.15) we conclude that (2.12) holds. In particular, kr

k

(δ)k = kbk/σ

k

(δ), which together with (2.6) implies (2.13).

Remark 2.7. As σ

m−1

(δ)/σ

N−`+1

= ||r

N−`+1

(δ)||/||r

m−1

(δ)||, we see that the decay rates in Corollary 2.4 and Corollary 2.5 are linked to the convergence of the GMRES algorithm. If the GMRES algorithm converges faster, the decay pattern becomes more pronounced.

Remark 2.8. By taking norms in (2.12), we see that for the relative GMRES residual kr

k

(δ)k

kr

0

(δ)k = 1

σ

k

(δ) = min

p∈Pk

kp(A)bk kbk|p(δ)| .

More generally, for the decay rates in Corollary 2.4 and Corollary 2.5 for ` ≤ k − m + 2, we

have σ

m−1

(δ)

σ

k−`+1

(δ) ≤ min

p∈Pk+2−m−`

kp(A)w

m−1

(δ)k

kw

m−1

(δ)k |p(δ)| , (2.17) Let Ω ⊂ C be a simply connected and compact K-spectral set for A; that is, kπ(A)k ≤ K max

z∈Ω

|π(z)| for all polynomials π (and hence Λ(A) ⊂ Ω) with δ 6∈ Ω. Also, let ϕ be a map, mapping C \ Ω conformally onto the exterior of the closed unit disk. Then it can be shown that the right-hand side of (2.17) can be bounded above by |1/ϕ(δ)|

k+2−`−m

< 1 times a modest constant, see, e.g., [16, Chapter 6.11.2] for the case where Ω is an ellipse.

Example 2.9. As in Example 2.3, assume the matrix A is unitary. Given a polynomial p of degree k, its reversed polynomial p

is defined as

p

(z) := z

k

p(1/z).

The orthogonal polynomials q

k

(z) can be expressed as

q

k

(z) = q

k

(0) argmin

qmonic of degreek

kqk

A,v1

= q

k

(0) argmin

qof degreek

kqk

A,v1

|q

k

(0)| . (2.18)

Note that q

k

(0) is the leading coefficient of q

k

. As A is unitary, kqk

A,v1

= kq

k

A,v1

. Hence, (2.18) can be rewritten as

q

k

(z) = q

k

(0)argmin

qof degreek

kq

k

A,v1

|q

(0)|

Therefore,

q

k

(z) = q

k

(0)argmin

q∈P

k,q(0)6=0

kqk

A,v1

|q(0)| . (2.19)

Because of (2.4), q

k

(0) ≥ 0. Also, kq

k

k

A,v1

= 1. Combining this with (2.15) and (2.12), (2.19) yields

q

k

(z) = 1 σ

k

(0)

k

X

j=0

q

j

(0)q

j

(z), (2.20)

Therefore, the kth normalized GMRES residual w

k

(0) of a unitary matrix A can be expressed as

w

k

(0) = q

k

(A)v

1

. (2.21)

(10)

2.4 A progressive GMRES residual formula

By (2.13) we have that the kth normalized GMRES residual satisfies w

k

(δ) = 1

σ

k

(δ)

k

X

i=0

q

i

(δ)v

i+1

= σ

k−1

(δ) σ

k

(δ)

1 σ

k−1

(δ)

k−1

X

i=0

q

i

(δ)v

i+1

+ q

k

(δ) σ

k

(δ) v

k+1

= σ

k−1

(δ)

σ

k

(δ) w

k−1

(δ) + q

k

(δ) σ

k

(δ) v

k+1

,

= s

k

(δ)w

k−1

(δ) + (−1)

k

c

k

(δ)v

k+1

, (2.22) the last equality following from (2.8). This demonstrates the existence of an updating formula to compute the residual vectors progressively. This formula can also be derived by means of the QR-factorization of H

k

− δI

k

, see Proposition 2.2 and [16, Subsection 6.5.3]. The next result shows that it is not necessary to compute such a factorization for obtaining s

k

(δ) and c

k

(δ), if one is willing to compute the additional scalar product (2.23).

Proposition 2.10. Define τ

k

(δ) = e

k

Q

k

(δ)

(H

k

− δI

k

)e

k

. Then

τ

k

(δ) = w

k−1

(δ)

(A − δI

n

)v

k

(−1)

k−1

, (2.23) and

s

k

(δ) = h

k+1,k

q

h

2k+1,k

+ |τ

k

(δ)|

2

and c

k

(δ) = τ

k

(δ) q

h

2k+1,k

+ |τ

k

(δ)|

2

. (2.24)

Proof. From Proposition 2.2 and (2.13) it follows that

V

k

Q

k

(δ)e

k

= V

k

 q

0

(δ)

. . . q

k−1

(δ)

(−1)

k−1

σ

k−1

(δ)

= (−1)

k−1

σ

k−1

(δ)

k

X

i=1

v

i

q

i−1

(δ) = (−1)

k−1

w

k−1

(δ). (2.25) Therefore,

τ

k

(δ) = e

k

Q

k

(δ)

(H

k

− δI

k

)e

k

= e

k

Q

k

(δ)

V

k

V

k

(H

k

− δI

k

)e

k

= w

k−1

(δ)

(Av

k

− h

k+1,k

v

k+1

− δv

k

)(−1)

k−1

= w

k−1

(δ)

(A − δI

n

)v

k

(−1)

k−1

, establishing (2.23). We now claim that

w

k

(δ) ⊥ (A − δI

n

)v

j

, for 1 ≤ j ≤ k, (2.26) which is a direct consequence of (2.14):

r

k

(δ) = r

0

(δ) − u

k

with u

k

= argmin

u∈K

kr

0

(δ) − uk, where K = (A − δI

n

)span{v

1

, . . . , v

k

}.

It remains to prove (2.24). Taking inner products with (A − δI

n

)v

k

in all terms of (2.22) and making use of (2.23) and (2.26), results in

0 = s

k

(δ)τ

k

(δ)(−1)

k−1

+ c

k

(δ)h

k+1,k

(−1)

k

. (2.27)

Together with s

2k

(δ) + |c

k

(δ)|

2

= 1, this yields (2.24).

(11)

Example 2.11. Let us return to the particular case of δ = 0 and unitary A as discussed in Examples 2.3 and 2.9. Inserting (2.6) and (2.21) in (2.22) and identifying the underlying polynomials gives

q

k

(z) = s

k

(0)q

k−1

(z) + (−1)

k

c

k

(0)q

k

(z). (2.28) Since q

k−1

(z) − q

k−1

(0) is z times a polynomial of degree < k − 1 and q

0

(z) = 1, we get from (2.23) that

τ

k

(0) = (−1)

k−1

hzq

k−1

, q

k−1

i

A,v1

= (−1)

k−1

q

k−1

(0)hzq

k−1

, q

0

i

A,v1

= (−1)

k−1

σ

k−1

(0)h

1,k

= c

k

(0)

and h

k+1,k

= s

k

(0), where we applied (2.20) and (2.9). Notice that this simplification for unitary A is in accordance with (2.24). Taking the star operation in (2.28) gives the second relation

q

k

(z) = s

k

(0)zq

k−1

(z) + (−1)

k

c

k

(0)q

k

(z). (2.29) We should mention that, in case of unitary A, the scalar product (2.2) can be written as a scalar product in terms of a (discrete) measure supported on the unit circle. Thus we have the whole theory of orthogonal polynomials on the unit circle, and in particular relations (2.28) and (2.29) are known as the Szeg˝ o recurrence relations, a coupled two-term recurrence for orthonormal polynomials on the unit circle [11, Formula 1.2-1.7].

3 The Arnoldi process for BM L-matrices

We will establish a fast variant of the Arnoldi process which is applicable to the class of BM L-matrices as described by formula (1.1). In §3.1 we will use the explicit representation of (H

k+1

− δI

k+1

)

−∗

in terms of orthogonal polynomials as stated in (2.11) to demonstrate that the underlying Hessenberg matrix has an upper-separable structure. The assumption on simple poles makes it possible to easily express generators for the upper-separable structure in polynomial language. In particular, the GMRES residual vectors, which can be expressed in terms of orthogonal polynomials by (2.13), are showing up. In §3.2 we will see that the orthonormal basis vectors (up to a correction term incorporating the low rank perturbation in (1.1)) can be written as a linear combination of previously computed basis vectors and GMRES residual vectors. In §3.3 the results from §3.2 and §2.4, enabling to compute the necessary GMRES residual vectors progressively, will be combined into a new Arnoldi iteration for BM L-matrices.

3.1 Structure formula for BM L-matrices

A rank structure revealing formula for BM L-matrices will be derived in terms of the orthog- onal polynomials q

k

defined in §2.1. This formula (3.3) will be the key for the design of short recurrence relations in the following subsection. We first show that the BM L-structure (1.1) is inherited by the underlying Hessenberg matrix.

Proposition 3.1. Let A be an invertible matrix satisfying relation (1.1), and denote by N ≤ n the Arnoldi termination index. Then H

N

is also a BM L-matrix with the same polynomials p, q and indices m e

1

= m

1

, m e

2

= m

2

and m e

3

≤ m

3

. More precisely,

H

N

= p(H

N

)q(H

N

)

−1

+ F

N

G

N

(3.1) where F

k

:= V

k

F and G

k

:= V

k

G.

Proof. By definition of N we have AV

N

= V

N

H

N

, i.e., the columns of V

N

span an invari- ant subspace of A. By recurrence on the degree one shows that π(A)V

N

= V

N

π(H

N

), or π(H

N

) = V

N

π(A)V

N

, for any polynomial π. Moreover, if π(A) is of full rank, then so is π(H

N

). In particular, since we assume q(A) to be invertible, so is the matrix q(H

N

). Hence, V

N

q(H

N

)

−1

= q(A)

−1

V

N

. This implies that

H

N

− p(H

N

)q(H

N

)

−1

= V

N

(A

− p(A)q(A)

−1

)V

N

= V

N

F G

V

N

,

as claimed in (3.1).

(12)

A combination with Corollary 2.5 gives us the following result.

Corollary 3.2. Let A be a matrix satisfying relation (1.1). Denote by N ≤ n the Arnoldi termination index, and m := max(0, m

1

−m

2

+1). Then H

N

∈ C

N×N

is (m, m

2

+m

3

)−upper- separable. More precisely,

((H

N

)

k,`

)

`=k+m,...,N

=

m2

X

j=1

d

j

q

k−1

(z

j

)

(H

N

− z

j

I

N

)

−∗

1,`

`=k+m,...,N

+

(G

N

F

N

)

k,`

`=k+m,...,N

, (3.2)

where z

1

, . . . , z

m2

denote the poles of the rational function p(z)/q(z).

Proof. The fact that H

N

is (m, m

2

+m

3

)−upper-separable is already known from [14]. Below we give an alternative, more constructive, proof leading to explicit generators for the low rank part of H

N

. By assumption, the rational function p/q in (1.1) has simple poles and thus has the partial fraction decomposition

p(z) q(z) =

m2

X

j=1

d

j

z − z

j

+ π(z),

for some constants z

j

, d

j

and a polynomial π of degree m

1

−m

2

(π = 0 if m

1

< m

2

). Replacing z by H

N

, taking adjoints and using (3.1) and (2.11) leads to

H

N

− G

N

F

N

− π(H

N

)

=

m2

X

j=1

d

j

(H

N

− z

j

I

N

)

−∗

= L

N

+

m2

X

j=1

d

j

q

0

(z

j

), . . . , q

N−1

(z

j

)

e

1

(H

N

− z

j

I

N

)

−∗

, (3.3) with a strictly lower triangular matrix L

N

. Finally, according to the upper Hessenberg struc- ture of H

N

, the matrix π(H

N

)

has zero entries on and above the mth diagonal, establishing the upper-separable structure and formula (3.2).

Remark 3.3. The assumption on simple poles is not necessary for H

N

to have an upper- separable structure. However, once we drop this constraint, it is not clear whether or not there exists a link between the generators of the low-rank structure and the GMRES residual vectors. We refer to [14] for more information on this topic.

Note that H

k

for k < N also has an upper-separable structure as it is a leading principal minor (submatrix) of H

N

. However, H

k

does not satisfy the same matrix equation as H

N

.

3.2 Short recurrence relations for BM L-matrices

We will now derive a short recurrence relation for BM L-matrices. To do so, the vector v

0k

:= Av

k

− V

k−m

G

k−m

F

v

k

, (3.4) is introduced, which is equal to Av

k

up to a correction term induced by the low rank pertur- bation in (1.1). In [4] it is shown that for nearly Hermitian A, i.e., p(z) = z and q(z) = 1 in (1.1), the Arnoldi vectors satisfy a three term recurrence relation. More specifically,

v

0k

= h

k−1,k

v

k−1

+ h

k,k

v

k

+ h

k+1

v

k+1,k

, (3.5)

where h

k−1,k

= v

k−1

v

k0

, h

k,k

= v

k

v

0k

and h

k+1,k

= v

k+1

v

0k

are entries of the underlying

Hessenberg matrix. We refer to [4] for a detailed discussion. For a general BM L-matrix,

Proposition 3.4 states that v

0k

is a linear combination of GMRES residual vectors and Arnoldi

vectors, including v

k+1

. As we will discuss below, this results in a short recurrence relation

which reduces to (3.5) in the specific case of nearly Hermitian matrices and which can be used

to compute the Arnoldi vectors in an efficient way.

Références

Documents relatifs

For finitely generated group that are linear over a field, a polynomial upper bound for these functions is established in [BM], and in the case of higher rank arithmetic groups,

Write down the transition matrix between the bases

Even though solutions develop large gradients and demand a large number of Fourier modes to resolve, the adaptive basis functions maintain accuracy of residual diffusivities at

The reader may consult [2] for a proof of Grothendieck's inequality and for supplementary information... Lindenstrauss and

Finally, for all the positive RI outcome groups, the MRIL was obtained at lower levels for noises with cen- tral frequency at the tinnitus frequency or at the hearing loss

Toute utili- sation commerciale ou impression systématique est constitutive d’une infraction pénale. Toute copie ou impression de ce fichier doit conte- nir la présente mention

In Tables 5.5 through 5.10 we provide the individual and total errors, the experimental rates of convergence, the a posteriori error estimators, and the effectivity indices for

First, the final ellipse used by Saad may not be the optimal Tchebychev ellipse for the mini- max solution, because Saad only finds the exact ellipse relative to the real part Re (\ r