• Aucun résultat trouvé

Scalability analysis for two dimensional chains of fixed size subdo-

1.4 Scalability of domain decomposition methods

1.4.1 Scalability analysis for two dimensional chains of fixed size subdo-

In this subsection we study the scalability properties of OSMs for a two dimensional chain of fixed size subdomains. For the one dimensional chain analysis, we refer the interested reader to [28, Section 4.1]. Let us considerL>0 andδ, 0<δ<L2, and define the grid points aj for j=1, . . . ,N+1 andbj for j=0, . . . ,N as shown in Figure 1.3. The j-th subdomain of the chain is a rectangle of dimensionΩj :=(aj,bj)×(0,bL), andΩ:= ∪Nj=1j. We are interested in the solution to

−∆u=f inΩ, u=g on∂Ω. (1.4.1)

We consider directly the error equation and we define the errorsenj :=uunj, whereunj are the iterates of the OSM. In the error form, the overlapping OSM with Robin transmission conditions with parameterpis given by

−∆enj =0 inΩj, enj(·, 0)=0,enj(·,L)b =0,

xenj(aj,·)−penj(aj,·)=xenj−11(aj,·)−penj−11(aj,·),

xenj(bj,·)+penj(bj,·)=xenj+11(bj,·)+penj+11(bj,·),

(1.4.2)

forj=2, . . . ,N−1, and

−∆ 1 = Ω1

To study the iteration, we use the Fourier expansionenj(x,y)=P

k∈K vnj(x,k) sin(k y) with K :=nπ

Lb,

bL , . . .o

. Inserting this expansion into (1.4.2) and (1.4.3), the Fourier coefficients vnj satisfy solution to (1.4.4) is given by

vnj(x,k)=Rn1(aj)

where

We are ready to prove the scalability of the OSM in the overlapping case.

Theorem 1.4.1. Recall(1.4.7)and defineϕ(k,δ,p) := |g3−pg1|+|g4−pg2|. The overlapping OSM(1.4.2)is scalable, in the sense that

ρ(T2DO (k,δ,p))≤ kT2DO(k,δ,p)k≤max

k max{ϕ(k,δ,p),kTe1(k,δ,p)k}<1, independently of N for every p≥0.

Proof. Because of the structure ofT2DO , the normkT2DO kis given by

By computing the derivative ofϕwith respect topwe find

∂ϕ

∂p =− 2ke2δk+2kL−2ke2δk

k2e4δk+2ke4δk+p2e4δke2kL+2k2e2δk−2e2δkp2ekL+(p−k)2 forp<k,

∂ϕ

∂p = 2ke2δk+2kL−2ke2δk

k2e4δk +2ke4δk+p2e4δke2kL+2p2e2δk −2e2δkk2ekL+(k−p)2 forp>k.

Analyzing the signs of these derivatives, we see thatϕ(k,δ,p) is strictly decreasing for p<kand it is strictly increasing forp>k, thus it reaches a minimum forp=k. Therefore the maximum ofϕ(k,δ,p) with respect to the variable p is obtained for p=0 and for p→ +∞:

ϕ(k,δ,p)≤max©

ϕ(k,δ, 0) , limp

→∞ϕ(k,δ,p)ª . Forp=0,δ>0 andL>0 we have

ϕ(k,δ,p=0)=e2δke2δk+ekLekL ekL+2δke−kL−2δk

= sinh(2δk)+sinh(kL)

sinh(kL) cosh(2δk)+sinh(2δk) cosh(kL)<1, and, under the same conditions,

plim→∞ϕ(k,δ,p)= sinh(2δk)+sinh(kL)

sinh(kL) cosh(2δk)+sinh(2δk) cosh(kL)=ϕ(k,δ, 0)<1.

Hence, it holds thatϕ(k,δ,p)ϕ(k,δ, 0)<1. We now focus onkTe1kandkTe2k. Notice thatkTe1k= kTe2kand

kTe1k=

¯

¯

¯

¯

¯

(k+p)ekL+(k−p)ekL (k+p)ek(L+2δ)+(k−p)e−k(L+2δ)

¯

¯

¯

¯

¯

=

¯

¯

¯

¯

kcosh(kL)−psinh(kL) kcosh(k(L+2δ))+psinh(k(L+2δ))

¯

¯

¯

¯<1.

In order to get a bound independently ofk, we observe that lim

k→∞ϕ(k,δ,p)= lim

k→∞kTe1k= 0 if δ>0. Therefore defining ¯ρ(δ) :=maxkmax{ϕ(k,δ,p),kTe1(k,δ,p)k}, we see that kT2DO k=max{ϕ,kTe1k,kTe2k}<ρ¯(δ)<1, for everyδ,p>0.

For the case without overlap, we need a further argument because forδ=0 bothρ(T2DO ) andkT2DOkare less than one for any finite frequencyk, but tend to one ask→ ∞. One can therefore construct a situation where the method would not be scalable as follows:

suppose we haveN subdomains, and on the j-th subdomain we choose as initial guess e0j the j-th frequencye0j =eˆ0jsin(jπ

Lby). Then the convergence of the method is deter-mined by the frequency which maximizesρ(T2DO (k)). When the number of subdomains N becomes large, this maximum is attained for the largest frequencykN =N π

L yb since

ρ(T2DO(k))→1 ask→ ∞. Thus, every time we add a subdomain to the chain with a new initial condition on the interfaceN+1 according to our rule, the convergence rate of the method deteriorates fromρ(T2DO (Nπ

Lb)) toρ(T2DO ((N+1)π

Lb)) and the scalability property is lost. Theorem 1.4.2 gives however a sufficient condition such that the OSM is weakly scal-able also without overlap, and to see this we introduce the vectorenwithenk = krn(k)k

wherern(k) contains the Robin traces at the interfaces of thek-th Fourier mode.

Theorem 1.4.2. Given a toleranceTol, and supposing there exists ak that does not depend˜ on N such that e0k<Tolfor every k>k, then the OSM without overlap,˜ δ=0, and p>0is weakly scalable.

Proof. Suppose that the initial guess satisfieske0k>Tol, since otherwise there is noth-ing to prove. Then, due to the hypothesis, we have that maxπ

Lkk˜e0k>Tol. We now show that the method contracts with aρindependent of the number of subdomains up to the toleranceTol, and therefore we have scalability. Indeed, for everyksuch thatπLkk˜

enk= krn(k)k≤ kT2DO(k)kkrn−1(k)k≤ kT2DO ( ¯k)kkrn−1(k)k= kT2DO ( ¯k)kekn−1, wherekT2DO( ¯k)k=maxπ

L≤k≤k˜kT2DO (k)k<1 becausekT2DO (k)kis strictly less than 1 for every finitek. Now fork>k˜,

enk = krn(k)k≤ kT2DO(k)kkrn1(k)k≤ krn1(k)k=enk1,

sincekT2DO (k)k≤1. Therefore we observe that the method does not increase the error for the frequenciesk>k˜while it contracts for the other frequencies with a contraction factor of at least ¯ρ= kT2DO ( ¯k)k<1. Hence, as long askenk>Tol, we havekenkρ¯nke0k

with ¯ρindependent ofN.

The technical assumption in Theorem 1.4.2 on the frequency content of the initial error is not restrictive, since in a numerical implementation we have a maximum frequency kmax which can be represented by the grid. Choosing ˜k=kmax, the hypothesis of The-orem 1.4.2 is verified. Note also that without overlap,δ=0, we have thatkT2DOk=1 forp=0 orp→ ∞. Therefore we can not conclude that the method is scalable in these two cases. Forp=0, the OSM exchanges only partial derivatives information on the in-terface. Forp→ ∞, we obtain the classical Schwarz algorithm and it is well known that without overlap (δ=0), the method does not converge. We finally show the behaviour of p7→ kT2DO (k,δ,p)kfor a fixed pair (δ,k) in Figure 1.4. According to the proof of Theorem 1.4.1, the minimum of the functionp7→ϕ(k,δ,p) is located atp=k. Even though it is a minimum forϕ(k,δ,p) and not necessarily forkT2DO (k,δ,p)korρ(T2DO ), we might de-duce from Figure 1.4 that in order to eliminate thek-th frequency, a good choice would be to setp:=kin the OSM. For the Laplace equation, it has been shown for two subdomains that settingp:=kleads to a vanishing convergence factorρ(k) for the frequencyk[74].

In the case of many subdomains, a similar result has not been proved yet, but Figure 1.4 indicates that it might hold as well.

0 20 40 60 80 100 p

0 0.005

0.01 0.015 0.02

kTO 2D)k

Figure 1.4: Infinity norm of the iteration matrixT2DO as a function ofpforL=1,Lb=1,δ= 0.1,k=20,N=50.

a0 a1 · · · aj−1 aj · · · aN−1 aN x

1 · · ·j · · ·N

Lb

L L L

Figure 1.5: Nonoverlapping domain decomposition in two dimensions. Notice thataj= j L.

1.4.1.2 Scalability analysis for the Dirichlet-Neumann method

We now consider a two dimensional problem decomposed into nonoverlapping subdo-mains as shown in Figure 1.5. The error form of the parallel Dirichlet Neumann method (PDNM) is given by

−∆enj =0 inΩj, enj(·, 0)=0,enj(·,bL)=0,

enj(aj,·)=(1−θ)en−1j (aj,·)+θen−1j+1(aj,·),

xenj(aj1,·)=(1−µ)xen−1j (aj1,·)+µ∂xen−1j−1(aj1,·), forj=2, . . . ,N−1, and

−∆en1=0 inΩ1, e1n(·, 0)=0,en1(·,bL)=0, e1n(a0,·)=0,

e1n(a1,·)=(1−θ)en−11 (a1,·)+θe2n−1(a1,·),

and

where (1.4.10) may be rewritten as

en=T2DD Nen−1, infinity-norm is not suitable to bound the spectral radius and conclude convergence and scal-ability. Nevertheless in Theorem 1.4.3, we prove scalability of the PDNM under certain assumptions on the parametersµ,θand using similarity arguments..

Theorem 1.4.3. Denote by kminthe minimum frequency and defineα(x) :=1/ cosh(x). If θ=µ, then

minL)2, which implies that the PDNM is convergent and scalable.

We show in Figure 1.6 the function ¯ρ(µ) for the caseLb=1, that iskmin=π. The proof of Theorem 1.4.3 relies on the following lemma.

Lemma 1.4.4. Letα(x) :=1/ cosh(x). Then for any x∈(0,∞)such thatcosh(x)>2it holds

0 0.2 0.4 0.6 0.8 1

We are now ready to prove Theorem 1.4.3.

Proof. Ifµ=θ, the matrixT2DD Nhas the structure

whereB,eB,b B∈R2×2. We introduce an invertible block diagonal matrix

Notice that

Toff>Toff=diag Ã

0, 0,4µ2

γ22 , . . . ,4µ2 γ22 , 0, 0

! , and henceq

ρ(Toff>Toff)=2γµ2. Now, we focus on the termρ(Tdiag> Tdiag). The block diagonal structure ofTdiag> Tdiagallows us to write

ρ(Tdiag> Tdiag)= q

max©

ρ(Ce>Ce),ρ(Cb>Cb),ρ(C>C)ª

. (1.4.13)

The evaluation of the spectral radiiρ(Ce>Ce),ρ(Cb>Cb), andρ(C>C) leads to the analysis of cumbersome formulas, and we thus bound instead the spectral radii by the correspond-ing infinity-norms. To do so, settcorrespond-ingde1:=γ1andde2:=2, we obtain

ρ(Ce>C)e =ρ(GeBe>Ge−1Ge−1BeG)e ≤ kGeBe>Ge−1Ge−1BeGke =2µ2−2µ+1.

Next, we setdb1:=γ2anddb2:=1and get

ρ(C>C)=ρ(GBb >Gb−1Gb−1BG)b ≤ kGBe >Ge−1Ge−1BGek

=max (

1−µ, 1−µ+µ2(e−kLekL)4 (e−kL+ekL)4

)

≤1−µ+µ2,

where the fact that(e(e−kLkL−ekL)4

+ekL)4 ≤1 for anykis used. Now, a direct calculation shows that 2µ2−2µ+1≤2µ2−2µ+1+ 4µ(1−µ)

(ekminL+e−kminL)2≤1−µ2, for anyµ∈(0, 1). Therefore, we obtain

kTdiagk2=ρ(Tdiag> Tdiag)≤ q

1−µ+µ2. Recalling (1.4.12) and (1.4.13), we conclude that

ρ(T2DD N)≤ kTdiagk2+ kToffk2≤ q

1−µ+µ2+2µ γ2

≤ q

1−µ+µ2+ 2µ

(ekminL+e−kminL)=: ¯ρ(µ),

which is the first statement of the theorem. The second part follows now from Lemma 1.4.4 by observing that if ¯ρ(µ)<1, thenρ(T2DD N)≤ρ¯(µ)<1 where ¯ρ(µ) is independent of N.

1.4.1.3 Scalability analysis for the Neumann-Neumann method

Finally we study the convergence of the Neumann-Neumann method (NNM). For our model problem, the error form for the NNM is the following: first solve

−∆enj =0 inΩj, enj(·, 0)=0,enj(·,L)=0, enj(aj1,·)=Dnj−1,enj(aj,·)=Dnj, forj=2, . . . ,N−1 and

−∆en1=0 inΩ1, en1(·, 0)=0,en1(·,L)=0, en1(a0,·)=0,en1(a1,·)=D1n,

−∆enN=0 inΩN, enN(·, 0)=0,enN(·,L)=0, enN(aN1,·)=DnN−1,enN(aN,·)=0, then solve

−∆ψnj =0 inΩj,

xψnj(·, 0)=0,ψnj(·,L)=0,

xψnj(aj−1,·)=xenj(aj−1,·)−xenj1(aj−1,·),

xψnj(aj,·)=xenj(aj,·)−xenj+1(aj,·), forj=2, . . . ,N−1 and

−∆ψn1=0 inΩ1,

ψn1(·, 0)=0,ψn1(·,L)=0,ψn1(a0,·)=0,

xψn1(a1,·)=xe1n(a1,·)−xen2(a1,·),

and −∆ψnN=0 inΩN,

ψnN(·, 0)=0,ψnN(·,L)=0,ψnN(aN,·)=0,

xψnN(aN−1,·)=xenN(aN−1,·)−xenN1(aN−1,·), and finally set

Dnj+1:=Dnj −ϑ(ψnj+1(aj,·)+ψnj(aj,·)), (1.4.14) forj=1,· · ·,N−1, whereϑ>0. We expand bothenj andψnj in Fourier series ,

enj(x,y)= X m=1

vnj(x,k) sin(k y), ψnj(x,y)= X m=1

wnj(x,k) sin(k y), wherek∈K. The Fourier coefficientsvnj(x,k) andwnj(x,k) solve the problems

k2vnjxxvnj =0 in (aj−1,aj), vnj(aj−1,k)=Dnj1,

vnj(aj,k)=Dnj,

k2wnj −∂xxwnj =0 in (aj−1,aj),

xwnj(aj−1,k)=xvnj(aj−1,k)−∂xvnj1(aj−1,k),

xwnj(aj,k)=xvnj(aj,k)xvnj+1(aj,k),

forj=2, . . . ,N−1, and which is used to solve the problems inwnj, and we obtain

wnj(x,k)= 1

We definee = D1,D2,· · ·,DN1 , and write equations (1.4.15)-(1.4.16) asen+1=T2D e , where the iteration matrixT2DN Nis given by

T2DN N=

Proof. The infinity-norm ofT2DN Nis given by kT2DN Nk=max©

1. Therefore, we get kT2DN Nk=max holds, meaning thatγ1>2, and since the mapk7→γ1=2 sinh(kL) is strictly increasing in k, it suffices thatγ1>2 is satisfied for justk=π

Lb. Hence the condition becomes sinh(kL)>

1 or equivalentlykL>arcsinh(1)=ln(1+p

2), which concludes the proof.

1.4.1.4 Numerical results

We close this subsection with a numerical experiment. We start with a random initial guess and we apply the different methods to solve (1.4.1) withf =g=0. We set the ge-ometric parameters equal tobL=L=1 and we discretize each subdomain square with Nh=100 interior unknowns. For the overlapping OSM we chooseδ=10h, whereh=Nh1+1 and we setp=π. For the PDNM, we setθ=λ=12 while for the PNNMθ=14. In Table 1.1,

N 10 20 30 40 50

OSM 13 13 13 13 13

PDNM 70 70 70 70 70

PNNM 6 6 6 6 6

Table 1.1: Number of iterations to reach convergence as the number of subdomainsN increases.

we report the number of iterations to reach convergence with a tolerance of Tol :=10−12 for the different methods as the numberN of the subdomains increases. We can observe that every method requires a constant number of iterations to reach convergence. There-fore, these numerical experiments are in agreement with the theoretical results presented in this subsection. According to Table 1.1, it seems that the PNNM is the fastest method.

However, we remark that each iteration of the PNNM requires two subdomain solves, so its cost is comparable with the OSM. Moreover, the PNNM is extremely sensitive on the choice ofθ, see [26].