• Aucun résultat trouvé

Convergence of a splitting inertial proximal method for monotone operators

N/A
N/A
Protected

Academic year: 2021

Partager "Convergence of a splitting inertial proximal method for monotone operators"

Copied!
10
0
0

Texte intégral

(1)

HAL Id: hal-00780778

https://hal.univ-antilles.fr/hal-00780778

Submitted on 24 Jan 2013

HAL is a multi-disciplinary open access archive for the deposit and dissemination of sci- entific research documents, whether they are pub- lished or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.

Convergence of a splitting inertial proximal method for monotone operators

Abdellatif Moudafi, M. Oliny

To cite this version:

Abdellatif Moudafi, M. Oliny. Convergence of a splitting inertial proximal method for monotone operators. Journal of Computational and Applied Mathematics, Elsevier, 2003, 155 (2), pp.447-454.

�10.1016/S0377-0427(02)00906-8�. �hal-00780778�

(2)

proximal method for monotone operators

A. Moudafi and M. Oliny

Univeristé Antilles Guyane, DSI-GRIMAAG 97200 Schoelcher, Martinique, France.

abdellatif.mouda@martinique.univ-ag.fr

Abstract.

A forward-backward inertial procedure for solving the problem of nding a zero of the sum of two maximal monotone operators is proposed and its convergence is established under a cocoercivity condition with respect to the solution set.

Keywords.

monotone operators, elargements, proximal point algorithm, cocoerciv- ity, splitting algorithm, projection, convergence.

1 Introduction and preliminaries

The theory of maximal monotone operators has emerged as an eective and powerful tool for studying a wide class of unrelated problems arising in various branches of social, physical, engineering, pure and applied sciences in unied and general framework. In recent years, much attention has been given to de- velop ecient and implementable numerical methods including the projection method and its variant forms, auxiliary problem principle, proximal-point al- gorithm and descent framework for solving variational inequalities and related optimization problems. It is well known that the projection method and its variant forms cannot be used to suggest and analyze iterative methods for solv- ing variational inequalities due to the presence of the nonlinear term. This fact motivated to develop another technique, which involves the use of the resol- vent operator associated with maximal monotone operators, the origin of which can be traced back to Martinet [13] in the context of convex minimization and Rockafellar [20] in the general setting of maximal monotone operators. The resulting method, namely the proximal point algorithm has been extended and generalized in dierent directions by using novel and innovative techniques and ideas, both for their own sake and for their applications relying, for example, on Bregman distance. Since, in general, it is dicult to evaluate the resolvent op- erator. One alternative is to decompose the given operator into the sum of two

Accepted for publication in the Journal of Computation and Applied Mathematics

1

(3)

2

A.Moudaand M.Oliny

(or more) maximal monotone operators whose resolvent are easier to evaluate than the resolvent of the original one. Such a method is known as the operator splitting method. This can lead to the development of very ecient methods, since one can treat each part of the original operator independently. The oper- ator splitting methods and related techniques have been analyzed and studied by many authors including Eckstein and Bertsekas [8], Chen and Rockafellar [6], Zhu and Marcotte [27], P. Tseng [25] and Mouda and Théra [15]. For an excellent account of the splitting methods, see [7]. Here, we use the resolvent operator technique to suggest a forward-backward splitting method for solving the problem of nding a zero of the sum of two maximalmonotone operators. It is worth mentioning that if the nonlinear term involving the variational inequal- ities is the indicator function of a closed convex set in a Hilbert space, then the resolvent operator is equal to the projection operator and we recover a method proposed by A.S. Antipin [3]. Our result extends and generalizes the previously known results.

In this paper we will focus our attention on the classical problem of nding a zero of the sum of two maximal monotone operators

A

and

B

on a real Hilbert space

H

:

nd

x2H

such that

(A+B)(x)30:

(1.1) This is a well-known problem which includes, as special cases, optimization and min-max problems, complementarity problems, and variational inequalities.

One of the fondamental approaches to solving (1.1), where

B

is univoque, is the forward-backward method, which generates the next iterates

xk +1

by solving the subproblem

02

k

A(x)+(x?x

k +

k B(x

k

));

(1.2)

where

xk

is the current iterate and

k

is a regularization parameter. The lit- erature on this subject is vast (see

[7]

and references therein). Actually, this method was proposed by Lions and Mercier [12], by Passty [18] and, in a dual form for convex programming, by Han and Lou [10]. In the case where

A

is the normal cone of a nonempty closed convex set, this method reduces to a projec- tion method proposed by Sibony [23] for monotone variational inequalities and, in the further case where

B

is the gradient of a dierentiable convex function, it amounts to a gradient projection method of Goldstein and of Levintin and Polyak [5]. This method was largely analyzed by Mercier [14] and Gabay [9].

They namely showed that if

B

is cocoercive with modulus

>0

, then the iter- ates

xk

converge weakly to a solution on condition that

k

is constant and less than

2

. The case where

k

is noconstant was dealt with among others in [6, 8, 15, 25]

Recently, an inertial proximal algorithm was proposed by Alvarez in the con-

text of convex minimization in [1]. Afterwards, Attouch and Alvarez considered

its extension to maximal monotone operators

[2]

. Relying on this method, we

propose a splitting procedure which works as follows. Given

xk ?1;xk2H

and

(4)

two parameters

k2[0;1[

and

k >0

, nd

xk +12H

such that

k A(x

k +1 )+x

k +1

?x

k

?

k (x

k

?x

k ?1 )+

k B(x

k

)30:

(1.3) When

B =0

, the inspiration for (1.3) comes from the implicit discretization of the dierential system of the second-order in time, namely

d 2

x

dt 2

(t)+ dx

dt

(t)+A(x(t))30

a.e.

t0;

(1.4) where

>0

is a damping or a friction parameter.

When

H = IR2

,

A

is the gradient of a dierentiable function,

(1:4)

is a sim- plied version of the dierential system which describes the motion of a heavy ball rolling over the graph of

f

and which keeps rolling under its own inertia until stopped by friction at a critical point of

f

(see [4]). This nonlinear os- cillator with damping has been considered by several authors proving dierent results and

=

or identifying situations in which the rate of convergence of

(1:4)

or its discrete versions is better than those of the rst-order steepest descent method see (

[1;11;19]

). Roughtly speaking the second-order nature of

(1:3)

(re- spectively

(1:4)

) may be exploited in some situations in order to accelerate the convergence of the sequence of

(1:3)

(respectively the trajectories of

(1:4)

), see [11] where numerical simulations comparing the behavior of the standard proxi- mal algorithm, the gradient method and the inertial proximal one are presented ( for the continuous version see for example [4]).

For developing implementable computational techniques, it is of particular im- portance to treat the case when (1.3) is solved approximately. Before introduc- ing our approximate method, let us recall the following concepts which are of common use in the context of convex and nonlinear analysis. Troughout,

H

is a real Hilbert space,

h;i

denotes the associated scalar product and

jj

stands for the corresponding norm. An operator is said to be monotone if

hu?v ;x?y i0

whenever

u2T(x);v2T(y ):

It is said to be maximalmonotone if, in addition, the graph,

f(x;y )2H H:y2

T(x)g

, is not properly contained in the graph of any other monotone operator.

It is well-known that for each

x 2H

and

>0

there is a unique

z 2H

such that

x2(I+T)z

. The single-valued operator

JT :=(I +T)?1

is called the resolvent of

T

of parameter

. It is a nonexpansive mapping which is everywhere dened and satises:

z=JTz

, if and only if,

02Tz

. Let us also recall a notion which is clearly inspired by the approximate subdierential. In [21, 22], Iusem, Burachik and Svaiter dened

T"(x)

, an

"

-enlargement of a monotone operator

T

, as

T

"

(x):=fv2H;hu?v ;y?xi?" 8y ;u2T(y )g;

(1.5) where

"0

. Since

T

is assumed to be maximal monotone,

T0(x)=T(x)

, for any

x

. Furthermore, directly from the denition it follows that

0" " )T

"

1

(x)T

"

2

(x):

(5)

4

A.Moudaand M.Oliny

Thus

T"

is an enlargement of

T

. The use of elements in

T"

instead of

T

allows an extra degree of freedom, which is very useful in various applications. On the other hand, setting

" = 0

one retrieves the original operator

T

, so that the classical method can be also treated. For all these reasons, we consider the following scheme: nd

xk +12H

such that

k A

"k

(x

k +1 )+x

k +1

?y

k +

k B(x

k

)30:

(1.6)

where

yk :=xk+k(xk?xk ?1);k;k;"k

are nonnegative real numbers.

If

A

is the subdierential of the indicator function of a closed convex set

C

, then (1.1) reduces to the classical variational inequality

hB(x);y?xi0 8y 2C ;

(1.7) and the resolvent operator is nothing but the projection operator. Moreover, in the case where

"k =08k

and

B

is the gradient of a function

f

, (1.7) reduces in turn to the constrained minimization problem Min

x2Cf(x)

and we recover a method proposed by Antipin in [3], namely

x

k +1

=

proj

C(xk?rf(xk)+(xk?xk ?1)):

Anathor interesting case is obtained by taking

B=0

and

A=@f

,

@f

stands for the subdierential of a proper convex lower-semicontinuous function

f : H!

IR[f+1g

. Indeed,

@f

is well-known to be a maximal monotone operator and problem (1.1) reduces to the one of nding a minimizer of the function

f

. In

[1]

, Alvarez proposed the following approximate inertial proximal method:

k

@

"

k f(x

k +1 )+x

k +1

?x

k

?

k (x

k

?x

k ?1

)30;

(1.8)

where

@"kf

is the approximate subdierential of

f

. Since in the case

A=@f

the enlargement given in (1.5) is larger than the the approxiamte subdierential, i.e.

@"f (@f)"

(see [21, 22]), we can write

@"kf(xk +1)(@f)"k(xk +1);

which leads to

k (@f)

"

k

(x

k +1 )+x

k +1

?x

k

?

k (x

k

?x

k ?1

)30;

(1.9)

which is a particular case of the method proposed in this paper with

A=@f

and

B=0

.

In the sequel, we will need a cocoercivity condition with respect to the solution set,

S:=(A+B)?1(0)

, namely

hB(x)?B(y );x?y ijB(x)?B(y )j 2

8x2H8y2S;

being a positive real number. This condition is standard in the literature and

is typically needed to establish weak convergence (see for example [8], [9], [15],

[27]).

(6)

2 The main results

To begin with let us recall, for the convenience of the reader, a well-known result on weak convergence.

Lemma 2.1 Opial Let

H

be a Hilbert space and

fxkg

a sequence such that there exists a nonempty set

SH

verifying:

For every

x2S

, lim

k !+1jxk?xj

exists.

If

x

weakly converges to

x2H

for a subsequence

!

+

1

, then

x2S

. Then, there exists ~

x2S

such that

fxkg

weakly converges to

x

~ in

H

.

We are now able to give our main result.

Theorem 2.1 Let

fxkg H

be a sequence generated by (1.6), where

A;B

are two maximal monotone operators with

B

-cocoercive and suppose that the parameters

k;k

and

"k

satisfy:

1.

9" 9>

0 such that

8k2IN;k

2

?"

. 2.

92

[0

;

1[ such that

8k2IN;

0

k

. 3.

P+1k =1"k<

+

1

.

If the following condition holds

+1

X

k =1

k jx

k

?x

k ?1 j

2

<

+

1;

(2.10)

then, there exists

x

2S

such that

fxkg

weakly converges to

x

as

k!

+

1

.

Proof . Fix

x2S

=

T?1

(0) and set

'k

=

12jx?xkj2

. We have

'

k

?'

k +1

=

12jxk +1?xkj2

+

hxk +1?yk;x?xk +1i

+

khxk?xk ?1;x?xk +1i;

(2.11) where

yk

:=

xk

+

k

(

xk?xk ?1

). Since

?xk +1

+

yk?kB

(

xk

)

2kA"k

(

xk +1

) and

?kB

(

x

)

2kA

(

x

), from denition (1.5) it follows that

hx

k +1

?y

k

+

k

(

B

(

xk

)

?B

(

x

))

;x?xk +1i?k"k:

(2.12) Combining (2.11) and (2.12), we obtain

'

k

?'

k +1

1

2 jx

k +1

?x

k j

2

+

khB

(

xk

)

?B

(

x

)

;xk +1?xi

? hx ?x ;x ?xi? " :

(7)

6

A.Moudaand M.Oliny

By invoking the equality

h

x

k?

x

k ?1

;x

k +1?

x

i = h

x

k?

x

k ?1

;x

k?

x

i+h

x

k?

x

k ?1

;x

k +1?

x

ki

=

'

k?

'

k ?1+12j

x

k?

x

k ?1j2+h

x

k?

x

k ?1

;x

k +1?

x

ki

; it follows that

'

k +1?

'

k?

k(

'

k?

'

k ?1) ?12j

x

k +1?

x

kj2+

kh

x

k?

x

k ?1

;x

k +1?

x

ki

+ k

2

j

x

k?

x

k ?1j2?

kh

B

(

x

k)?

B

(

x

)

;x

k +1?

x

i+

k

"

k

: On the other hand, since B is cocoercive, we get

kh

B

(

x

k)?

B

(

x

)

;x

k +1?

x

i =

k(h

B

(

x

k)?

B

(

x

)

;x

k?

x

i+h

B

(

x

k)?

B

(

x

)

;x

k +1?

x

i)

k(

j

B

(

x

k)?

B

(

x

)j2+h

B

(

x

k)?

B

(

x

)

;x

k +1?

x

ki)

?

k

4

j

x

k +1?

x

kj2

:

From which infer, by setting

k :=1?2k

, the estimate (2.13) below

'

k +1?

'

k?

k(

'

k?

'

k ?1) ?12

kj

x

k +1?

x

kj2+

kh

x

k?

x

k ?1

;x

k +1?

x

ki

+

k

2

j

x

k?

x

k ?1j2+

k

"

k

?

1

2

kj

x

k +1?kk

y

kj2+2k2kj

x

k?

x

k ?1j2

+ k

2

j

x

k?

x

k ?1j2+

k

"

k

?

1

2

kj

x

k +1?

x

k?kk

(

x

k?

x

k ?1)j2

+ k

k

j

x

k?

x

k ?1j2+

k

"

k

:

By taking into account the fact that from the hypotheses

k

is bounded and by setting

k:=

'

k?

'

k ?1

and

k := 2k" j

x

k?

x

k ?1j2+

k

"

k

, we obtain

k +1

k

k+

k

k[

k]++

k

; where

[

t

]+:=

max

(

t;

0)

, and consequently

[

k +1]+

[

k]++

k

; with

2[0

;

1[

given by hypothesis 2.

The latter inequality yields

[

k +1]+

k[

1]++k ?1X

i=0

i

k ?i

;

and therefore

1

X

[

k +1] 1

1?

([

1]++X1

k)

;

(8)

which is nite thanks to hypothesis 3 and (2.10). Consider the sequence dened by

tk

:=

'k?Pki=1

[

i

]

+

. Since

'k

0 and

Pki=1

[

i

]

+<

+

1

, it follows that

tk

is bounded from below. But

t

k +1

=

'k +1?

[

k +1

]

+?Xk

i=1

[

i

]

+'k +1?'k +1

+

'k?Xk

i=1

[

i

]

+

=

tk;

so that

ftkg

is nonincreasing. We thus deduce that

ftkg

is convergent and so is

f'

k

g

. This show that the rst condition of Opial's lemma is satised.

On the other hand, from (2

:

13) we can write

1 2

kjxk +1?xk?kk

(

xk?xk ?1

)

j2?k +1

+

k

+

k:

By passing to the limit in the above estimate and by taking into account the conditions on the parameters and the fact that by hypothesis

jxk?xk ?1j!

0, we obtain

lim

k !+1 jx

k +1

?x

k

?

k

(

xk?xk ?1

)

j

= 0

:

Now let

x

be a weak cluster point of

fxkg

. There exists a subsequence

fxg

which converges weakly to

x

and satises, thanks to (1.6),

?

1

(

x+1?y

)+(

B

(

x+1

)

?B

(

x

))

2A" +1

(

x+1

)+

B

(

x+1

)

(

A

+

B

)

" +1

(

x+1

)

:

Passing to the limit, as

!

+

1

, using the fact that

B

is Lipschitz continuous and thanks to the properties of the enlargements ([22], proposition 3.4), we obtain that 0

2

(

A

+

B

)(

x

), that is

x2S

. Thus, the second condition of Opial's lemma is also satised, which completes the proof.

Condition (2

:

13) involves the iterates that are a priori unknown, in practice it is easy to enforce it by applying an appropriate on-line rule (for example, choosing

k

2

[0

;k

] with

k

:= min

f;(k jxk?xk?1j)1 2

g

. Furthermore, it is worth men- tioning that (2

:

13) is automatically satised in some special cases. For instance where assumption 2) of theorem 2.1 is replaced by

9 2

[0

;13

[

;8k 2 IN;

0

k

and the sequence

fkg

is nondecreasing (see [2], proposition 2.1).

Remark 2.1 An open problem is to develop a general theory to guide the choices of the parameters

k

and

k

.

Our result extends classical convergence results concerning the standard forward- backward method as well as theorem 6 of Antipin [3].

Acknowledgements

The authors are grateful to the two anonymous referees and to Professor P. B.

Monk for their valuable comments and remarks.

(9)

8

A.Moudaand M.Oliny

References

[1] F.

Alvarez

, On the minimizing property of a second ordre disipative system in Hilbert space , SIAM J. of Control and Optimization, Vol 38, N 4, (2000), p. 1102-1119.

[2] F.

Alvarez

and H.

Attouch

, An inertial proximal method for maximal monotone operators via discretization of a non linear oscillator with damp- ing . Set Valued Analysis, Vol. 9, (2001), p. 3-11.

[3] A. S.

Antipin

, Continuous and iterative Process with projection operators and projection-like operators, Voprosy Kibernetiki. Vychislitel'nye Voprosy Analiza Bol'shikh Sistem, AN SSSR, Scientic Counsel on the Complex Problem "Cybernetics", Moscow (1989), p. 5-43.

[4] H.

Attouch

, X.

Goudou

and P.

Redont

, The heavy ball with friction. I.

The continuous dynamical system , Commun.Contemp. Math 2, N 1 (2000), p. 1-34.

[5] D. P.

Bertsekas

, Constrained optimization and lagrange multiplier meth- ods , Academic Press, New York, 1982.

[6] H.-G.

Chen

and R.T.

Rockafellar

Convergence rates in forward- backward splitting , SIAM J. Optim., Vol. 7 (1997), p. 421-444.

[7] J.

Ekstein

, Splitting methods for monotone operators with application to parallel optimization. Dissertation, Massachussets Intitute of Technology 1989.

[8] J.

Ekstein

and D. P.

Bertsekas

, On the Douglas-Rachford splitting method and the proximal point algorithm for maximal monotone operators , Math. Programming, 55 (1992), p. 293-318.

[9] D.

Gabay

, Applications of the method of multipliers to variational inequal- ities , in Augmented Lagrangian Methods: Applications to the numerical solution of Boundary Value Problems, M. Fortin and R. Glowinski, eds, North-Holland, Amsterdam, (1983), p. 299-331.

[10] S. P.

Han

and G.

Lou

A parallel algorithm for a class of convex programs , SIAM J. Control Optim., 26 (1988), p. 345-355.

[11] F.

Jules

and P. E.

Maingé

Numerical approach to a stationary solution of a second order dissipative dynamical system , Optimization, 51, Issue 2 (2002), to appear.

[12] P. L.

Lions

and B.

Mercier

, Splitting algorithms for the sum of two non-

linear operators , SIAM J. Numer. Anal., 16 (1979), p.964-979.

(10)

[13] B. Martinet , Régularisation d'inéquations variationnelles par approxima- tions successives , Rev. Francaise Inf. Rech. Oper., (1970), p. 154-159.

[14] B. Mercier , Inéquations variationnelles de la mécanique , Pub. Math. Or- say, 80.01, Université de Paris-Sud, Orsay 1980.

[15] A. Moudafi and M. Théra , nding a zero of the sum of two maximal monotone operators , J. Optim. Theory Appl., 94 (1997), p.425-448.

[16] Z. Opial Weak convergence of the sequence of successive approximations for nonexpansive mappings , Bulletin of the American mathematicalSociety, Vol. 73, (1967), p. 591-597.

[17] J. M. Ortega and W. C. Rheinboldt Iterative solution of nonlinear equations in several variables , Acad. Press., 1970.

[18] G. B. Passty , Ergodic convergence to a zero of the sum of monotone op- erators in Hilbert spaces , J. Math. Anal. Appl., 72 (1979), p.383-390.

[19] B. T. Polyak Some methods of speeding up the convergence of iterative methods , Zh. Vychisl. Mat. Mat. Fiz. Vol4, (1964), p. 1-17.

[20] R.T. Rockafellar Monotone operator and the proximal point algorithm , SIAM J. Control. Opt., 14 (5), (1976), P. 877-898.

[21] B. F. Svaiter , R.S. Burachik , A.N. Iusem , Enlargement of maximal monotone operators with application to variational inequalities . Set-Valued Analysis, 5, (1997), p. 159-180.

[22] B. F. Svaiter , and R.S. Burachik ,

"

-Enlargements of maximal monotone operators in Banach spaces . Set-Valued Analysis, 7, (1999), p. 117-132.

[23] M. Sibony , méthodes itératives pour les équations et inéquations non linéaires de type monotone , Calcolo, 7 (1970), p.65-183.

[24] P. Tseng , Further applications of a splitting algorithm to decomposition in variational inequalities and convex programming , Math. Programming, 48, (1990), p. 249-263.

[25] P. Tseng , Applications of a splitting algorithm to decomposition in convex programming and variational inequalities , SIAM J. Control Optim., 29, (1991), p. 119-138.

[26] P. Tseng , A modied forward-backward splitting method for maximal monotone mapping , SIAM J. Control Optim., 38, (2000), p. 431-446.

[27] D. L. Zhu , and P. Marcotte , Co-coercivity and its role in the convergence

of iterative schemes for solving variational inequalities , SIAM J. Optim.,

Vol. 9, N 3 (1996), p. 714-726.

Références

Documents relatifs

proves that three apparently unrelated sets of convex cones, namely the class of linearly closed and convex cones of a finite-dimensional space, the class gathering all the cones

The hybrid proximal point algorithm introduced by Solodov and Svaiter allowing significant relaxation of the tolerance requirements imposed on the solution of proximal subproblems

More precisely, it is shown that this method weakly converges under natural assumptions and strongly converges provided that either the inverse of the involved operator is

Algorithm 2, with the sequence of update param- eter θ k described above, guarantees strict feasibility of the iterates and quadratic convergence of the proximity measure.. As

Then, when the non-smooth part is also partly smooth relative to a smooth submanifold, we establish finite identification of the latter and provide sharp local linear

In doing so, we put to the fore inertial proximal algorithms that converge for general monotone inclusions, and which, in the case of convex minimization, give fast convergence rates

In a Hilbert space setting, we study the asymptotic behavior, as time t goes to infinity, of the trajectories of a second-order differential equation governed by the

In [5], it was extended to non-smooth problems by allowing general proximity operators; The Barzilai-Borwein method [6] uses a specific choice of step-length t k motivated