• Aucun résultat trouvé

Average case analysis of greedy algorithms for optimisation problems on set systems

N/A
N/A
Protected

Academic year: 2021

Partager "Average case analysis of greedy algorithms for optimisation problems on set systems"

Copied!
32
0
0

Texte intégral

(1)

Laboratoire d'Analyse et Modélisation de Systèmes pour l'Aide à la Décision CNRS UMR 7024

CAHIER DU LAMSADE

125

Octobre 1994

Average case analysis of greedy algorithms for

optimisation problems on set systems

(2)

Contents

1 Introduction 1

2 Random set systems as random bipartite graphs 1

2.1 Random sets and random multisets . . . 1

2.2 Random set systems and bipartite graphs . . . 2

2.3 A particular class of set systems distributions . . . 2

2.4 Removing Vertices from the Incidence Graph . . . 3

3 Hitting Set and Set Cover 3 3.1 A Lower Bound for the Random Hitting Set Problem . . . 3

3.2 The GRECO algorithm . . . 4

3.3 Analysis of GRECO on random instances . . . 6

3.4 Proof of the convergence of of the sequence (τη)η∈IN∗ . . . 8

3.4.1 Explicit integration of the differential systems . . . 8

3.4.2 The computation of θη,h . . . 10

3.4.3 The signs of yI η,h,j . . . 10

3.4.4 Relations between θη,h and αη,h . . . 11

3.4.5 An upper bound of yη,h,j . . . 12

3.4.6 An upper bound of the sequence (τη)η∈IN∗ . . . 14

3.4.7 Variations with respect to α . . . 15

3.4.8 The convergence of the sequence (τη)η∈IN∗ . . . 25

4 Set packing and hypergraph independent set 26 4.1 The PACK algorithm . . . 26

4.2 Analysis of PACK on random instances . . . 26

4.2.1 Analysis of PACK on standard distribution . . . 26

4.2.2 The differential system simulating PACK for solving the hypergraph in-dependent set problem . . . 27

(3)

Analyse en moyenne de la performance des algorithmes gloutons

pour des probl`

emes d’optimisation sur des syst`

emes d’ensembles

R´esum´e

Nous pr´esentons un cadre g´en´eral pour l’analyse asymptotique d’algorithmes glou-tons pour plusieurs probl`emes d’optimisation sur des syst`emes al´eatoires d’ensem-bles lorsque le rapport entre la taille de l’ensemble de base et la taille du syst`eme d’ensembles reste constant. Les syst`emes d’ensembles sont engendr´es via des graphes biparties al´eatoires et les approximations des chaˆınes de Markov sont r´ealis´ees `a l’aide d’´equations diff´erentielles ordinaires.

mots-cl´es: Probl`eme NP-complet, algorithme polynomial approch´e, approximation en moyen-ne, couverture d’ensembles, transversal d’un hypergraphe, stable d’un hypergraphe

Average case analysis of greedy algorithms for optimisation

problems on set systems

Abstract

A general framework is presented for the asymptotic analysis of greedy algorithms for several optimisation problems such as hitting set, set cover, set packing, etc., applied on random set systems. The probability model used is specified by the size n of the ground set, the size m of the set system and the common distribution of each of its components. The asymptotic behaviour of each algorithm is studied when n and m tend to ∞, with mn a fixed constant. The main tools used are the generation of random families of sets via random bipartite graphs and the approximation of Markov chains with small steps by solutions of ordinary differential equations.

keywords: NP-complete problem, polynomial time approximation algorithm, average case ap-proximation, set covering, hitting set, hypergraph independent set

(4)

1

Introduction

There are many optimisation problems defined on families of sets such as hitting set, set cover, set packing, etc., which are known to be NP-complete, and for the (approximate) solution of which one must thus use either heuristics or greedy algorithms. The purpose of this paper is to present a general framework which permits the asymptotic analysis of greedy algorithms for several of these problems, including in particular those just mentioned.

This framework comprises two main parts:

(i) a representation of random set systems by using degree-constrained random bipartite graphs.

(ii) the introduction of a Markov chain on a certain state space for the analysis of each particular greedy algorithm.

It remains then to analyse the behaviour of this Markov chain. This behaviour can be approximated for problems of large size by the solution of an ordinary differential equation (or a system of such equations) which can be obtained in closed form.

We point out that Markov chains have been used previously, explicitely or implicitely, in the analysis of algorithms by various authors. Let us mention Fernandez de La Vega ([3]) and Frieze Radcliffe and Suen ([5]).

This paper should be considered mainly as a theoretical contribution to the analysis of greedy algorithms on set systems and, accordingly, we have not included numerical results. Such results would be of limited help here since in most cases (a basic obstruction here, comes from the fact that) we do not know the values of the optimum of the objective functions (the approximate values which may be obtained via the use of the first moment method seem to be rather inaccurate).

A preliminary version of this paper has been presented in the LATIN’92 conference ([4]). The plan of this paper is as follows. In § 2 we introduce the model of random set systems that we use and a useful connection with random bipartite graphs. In § 3, we study in details an algorithm for set covering and hitting set and in § 4 we show how this algorithm can be modified to be used for set packing and hypergraph independent set. Finally, in the appendix, we prove the convergence of the differential system describing the set covering algorithm.

2

Random set systems as random bipartite graphs

2.1 Random sets and random multisets

It will be convenient to work with families of random multisets rather than random sets. Moreover, we will view our families as ordered. In other words, we introduce the uniform measure µ on the set of ordered k-tuples from some finite ground set S and work with µm,

the m-fold product of µ defined on the set of ordered families of k-tuples of size m. Let us denote by ν = νm the more conventional mesure which gives the same weight to each

m-family of pairwise distinct k-sets. It is easy to see that when sampling on genuine k-sets with replacement we get m distinct sets with probability near to 1 if only m = o(nk/2).

(5)

clearly p = (n − 1)(n − 2) . . . (n − k + 1) nk−1 ≥ exp ( −k 2 2n ) .

Thus, setting λ = m/n, the probability pm of getting m genuine k-sets when sampling

according to µm satisfies pm ≥ exp {−(1/2)λk2}, and this together with the former result,

implies that, for a fixed λ, the µm measure of the genuine systems of m k-sets drawn from

a set of size n exceeds exp{−λk2}. Thus, again for a fixed λ, every event “almost sure”

relatively to the sequence of measures µm, m → ∞ is also almost sure relatively to the

more conventional measure νm and this property suffices for our purposes.

2.2 Random set systems and bipartite graphs

Let X and Y denote sets with the same cardinality. Let P = {X1, X2, . . . , Xm} denote

a partition of the set X into sets with sizes m1, m2, . . . , mm. Let P = (p1, p2, . . . , pm)

where pi denotes the number of classes of P of size i. Similarly, let Q = {Y1, Y2, . . . , Yn}

denote a partition of the set Y into sets with sizes d1, d2, . . . , dnand let Q = (q1, q2, . . . , qn)

be defined similarly as P .

Now, let Π denote a random pairing between the sets X and Y . If we contract each set of vertices Xito a single vertex and similarly each set Yj, we get a bipartite graph, say B, with

partitions M = (m1, m2, . . . , mm) and D = (d1, d2, . . . , dn). If we think of Q as a fixed

ground set, this pairing defines naturally a random set-system P = {X1, X2, . . . , Xm},

where we identify Xi with the set of classes Yj adjacent to it in Π.

Notice that this pairing procedure gives the same measure to all genuine set systems, in which no set contains more than once any element of the ground set and no two sets are equal. Indeed, any such system is given by precisely Qm

i=1mi!Qni=1dj! distinct pairings.

Henceforth, we have at our disposal a procedure for generating random systems of sets with fixed cardinalities, subject moreover to the condition that each element of the ground set belongs to a prescribed number of sets in the system.

2.3 A particular class of set systems distributions

We are in fact mostly interested in the following class of probability distributions on set systems. We postulate that the cardinalities of our random sets X1, X2, . . . , Xm are

independent random variables N1, N2, . . . , Nm with a common distribution F defined by

its individual probabilities {pj = Pr[N1 = j], j = 1, 2, . . .}. Moreover, we postulate that,

for each fixed j, the set Xi is, conditionally on Ni = j, uniformly distributed between the

(multi) subsets of [n] of cardinality j.

Thus, if we choose for F the distribution concentrated on some fixed integer k, we get a random k-uniform set system of size m. If we give a fixed value to the ratio m/n, say m/n = λ, then, by the law of large numbers, the number nh of vertices of degree h in

the corresponding incidence graph satisfies, for each integer h ≥ 0, nh/n → e−λλh/h! in

probability as n → ∞.

Now, if we want to generate a random set system according to some distribution in the class just defined, we are faced with two difficulties:

(6)

i) the degrees are unbounded,

ii) the partition of the incidence graph is not fixed.

In order to circumvent the first difficulty, we restrict our algorithms to work only on vertices with a bounded (but arbitrarily large) degree ∆ and we show that the size of the obtained solutions tends to a limit as ∆ → ∞.

The second difficulty is circumvented by conditioning on the degrees of the graph. Indeed, since the proportion of vertices of each fixed degree in each color class of the incidence graph B = Bn tends to its expectation when n → ∞ and, as it will be seen, the sizes

of the obtained solutions are continuous relatively to the ratios nh/n, we can obtain the

limits of these sizes by replacing these ratios by their expectations.

2.4 Removing Vertices from the Incidence Graph

Let us return to the graph B and let us look at what happens when we remove a ver-tex y randomly chosen from the vertices in Y of degree h in B, i.e, we remove the h corresponding vertices in V (P ) (then, all the sets containing y are removed). Since the pairing P between the sets X and Y is random, the subset of Y \ {y} whose vertices are adjacent to the removed sets in X is also random. It can then easily be checked that the new incidence graph is again random within the graphs which have the same degrees. We note that, since B may eventually be a multigraph, the number of removed sets may (exceptionally) be strictly smaller than h. We will of course take this fact into account in our calculations.

It will be seen that all the algorithms we consider, remove at each step either vertices from Y chosen on the basis of their degrees, or a vertex Xi chosen on the basis of the

size |Xi|, or both. This implies again that the successive incidence graphs which appear

are random within the graphs which have the same partition.

3

Hitting Set and Set Cover

3.1 A Lower Bound for the Random Hitting Set Problem

Let C denote a fixed subset of [n] of cardinality l. The probability that C intersects Ci

for some fixed i is 1 − [1 − (l/n)]k. Since the Ci’s are independent, the probability that

C intersects Ci for each i is p =

h

1 − [1 − (l/n)]kim. Hence, the expectation EYl of the

number Yl of sets of cardinality l which intersects each Ci is given by

n l ! 1 − 1 − l n !k  m . (1)

Setting m = λn and l = βn, we get log EYl ∼ n −λ log 1 1 − (1 − β)k + log 1 ββ(1 − β)1−β ! .

(7)

This implies EYl = o(1) whenever λ > λβ = log 1 ββ(1−β)1−β log1−(1−β)1 k . (2)

The following assertions can be easily deduced from this formula. Assertion 1. For any fixed k, β tends to 1 when λ tends to ∞.

Assertion 2. For any fixed λ, the ratio βk/ log k tends to 1 when k tends to ∞. Proof of assertion 1.

Putting γ = 1 − β, and using the inequalitynl≤ (ne/l)l, we get from (1), E(Yl)1/n ≤  1 − γkλ e γ !γ .

Suppose for a contradiction that γ ≥ γo > 0. Then, when λ → ∞, the first term on

the righthandside tends to 0 whereas one can easily check by taking derivatives that the second term is bounded above by e. Therefore, EYl → 0 as λ → ∞.

Proof of Assertion 2.

Let us prove first that we have β ≥ k−1log k(1 − o(1)). Inserting in (2) the value β =

(log k − 2 log log k)/k and using the inequality β log(β−1) ≤ 1, we get

λβ ≤ log 1−β1 log 1 1−e−βk ≤ β + 2β 2 e−βk − e−β2k2 ≤ 2k−1log k

k−1log2k(1 − o(1)) = o(1).

The proof of the reverse inequality is similar and is omitted.

3.2 The GRECO algorithm

For any family C of sets, a set H is a hitting set for C if H has a non-empty intersection with each element of this family and the minimum hitting set problem is that of finding a hitting set of minimum cardinality (see also [6]). We note that this problem is the dual (via the interchange of the two vertex sets of the incidence graph) of the set cover problem.

We note that recently, Lund and Yannakakis in [10] have proved that unless NP ⊆ DTIME[npoly log n] (conjecture weaker than P = NP but highly improbable), there is

no polynomial time algorithm approximating the optimal solution of these two problems with a ratio smaller than c log n for a constant c < 1/4 (this result is of course a “worst case” result). Consequently, average case studies of approximation algorithms for these problems are of high theoretical and practical importance. Let us mention that another algorithm due to Karp has been analysed with the same probability model ([7]).

Let the family of sets P = {X1, X2, . . . , Xm}, together with the corresponding partitions P

and Q, be defined as above via the random pairing Π. We consider the greedy algorithm 1 for finding a hitting set for P. Informally, our algorithm GRECO selects at each step

(8)

begin

split every set Yi with degree di > η into di new vertices of degree 1

for h ← η to 1 do while|X| 6= 0 do

choose randomly one of the Yi’s with maximum di;

delete Yi;

delete X = Xi1, . . . , Xih, the set of the classes in P, incident to Yi in Π;

delete any other vertex in Y adjacent to at least one of the classes of X; let X, Y, P, Q denote now the new configuration;

let Π denote the survived incidence graph od

od end;

Algorithm 1: GRECO algorithm. We suppose that after the splitting of the Yi

per-formed in the first line of the algorithm, the incidence graph Π is left untouched).

begin

γ ← ⌈n log k/k⌉; SC ← {x1, . . . , xγ}

for j ← 1 to m do

ifCj∩ SC = ∅ then add to SC an arbitrarily chosen element of Cj

od; end.

Algorithm 2. QUICK algorithm

one element chosen at random between the elements hitting the biggest number of sets which have not been hitted before. It depends on an integer parameter η.

Before proceeding to the analysis of this algorithm let us remark that when the product λk is high, the distribution of the degrees tends to uniformity. It turns out that in this case the much simpler algorithm QUICK (algorithm 2) provides asymptotically optimal solutions.

Let us proceed to the analysis of QUICK.

Plainly the size of the solution is bounded above by γ + µ, where µ denotes the number of elements of C which do not intersect the set {x1, . . . , xγ}. We have

Eµ = λn n−γ k  n k  ∼ λn 1 − log k k !k .

(9)

of its expectation. This implies, using Tchebicheff inequality and the familiar inequality 1 + x < ex, that Pr [µ ≤ λn/k] →

n→∞ 1.

Thus, again with probability tending to 1, the total size of the solution found by QUICK is bounded above by (n log k/k) + (λn/k).

Assertion 2 then implies that the approximation ratio of QUICK tends to 1 when k tends to ∞. Moreover, it can be shown that algorithm GRECO works better than QUICK; thus, assertion 2 provides also an approximation guarantee for GRECO when k tends to ∞.

3.3 Analysis of GRECO on random instances

Let us consider again a random configuration X, Y, P, Q, Π and the corresponding par-titions P and Q. We are interested here in a system of m random sets of size k, drawn from a ground set of size n and we assume that the ratio m/n is fixed, say m/n = λ, and we let m (and n) go to infinity. We must analyse separately the phases of the algorithm corresponding to the successive removals of the Q-vertices of degree η, η − 1, . . . , 1. Let h be a positive integer satisfying 1 ≤ h ≤ η and assume that, after having used vertices with degrees greater than h, there remain, for 1 ≤ j ≤ h, precisely mj Q-vertices of degree j.

Then, as it was pointed out before, the conditional distribution of the remaining incidence graph (conditioned by the previous steps of the algorithm) coincides with the “random pairing” distribution corresponding to the nj’s and mj’s (the nj’s being here all equal

to k).

Let us remove a (randomly chosen) Q-vertex of degree h. Then: • Pr[h new sets in P are captured] = 1 − o(1);

• the total decrease of the degrees of the remaining vertices has expectation h(k −1)−o(1) and the probability that the degree of any given Q-vertex is decreased by 1 is equal to [(h(k − 1) − o(1))j]/P

1≤i≤himi, where j denotes the degree of X;

• the decrease of the total size of the remaining Q-vertices has expectation h(k − 1) + o(1) and this implies that, setting S = h(k − 1)/P

1≤i≤himi, the expectations E∆mj of the

increments ∆mj satisfies

E∆mj = (j + 1)mj+1S − jmjS, 1 ≤ j ≤ h − 1,

E∆mh = −1 − hmhS.

Let Rη,h= (1/n)P1≤i≤himi denote the average degree of the Q vertices. Notice that the

total decrease of the degrees of these vertices in one step is h + h(k − 1) = hk. Standard results concerning the approximation of “small steps Markov chains” (see for instance Proposition 4.1 in [9]) imply, setting yη,h,j = mj/n and θ = t/n, that the yη,h,j’s are well

approximated as n → ∞ by the solution ~y of the following system of differential equations (where we use Newton’s notation for dy/dθ)

(Sη,h)    ˙yη,h,j(θ) = (j + 1)Rh(k−1) η,h−hkθyη,h,j+1(θ) − j h(k−1) Rη,h−hkθyη,h,j(θ), 1 ≤ j ≤ h − 1 ˙yη,h,h(θ) = −1 − hRh(k−1)η,h−hkθyη,h,h(θ)

in the sense that, sup

θ

|yη,h,j(θ) − (1/n)mj(θ)| → 0 as n → ∞, in probability for each

(10)

This concludes the description of the systems (Sη,h), 1 ≤ h ≤ η.

Let us mention that the use of three indices η, h and j which may seem luxurious to the reader, is in fact needed for the mathematical study of the above systems (see the appendix).

Let us emphasize now how we use these systems to analyse GRECO with parameter η. We integrate first (Sη,η) under the following initial conditions:

yI η,η,1(0) = e−λ+ ∞ X j=η+1 e−λλj (j − 1)! yIη,η,j(0) = e −λλj j! , 1 ≤ j ≤ η (3)

where λ = m/n is the average degree of the Q-vertices of Π.

Observe that for each j, the corresponding RHS of the above equalities is equal to the almost sure limit of mj(0)/n as n → ∞. Also, Rη,η =Pηj=1jyη,η,jI . We denote by θη,η the

first time at which the last coordinate becomes 0:

θη,η = min{θ > 0 : yη,η,η(θ) = 0}.

We set yI

η,η−1,j = yη,η,j(θη,η), 1 ≤ j ≤ η − 1, and Rη,η−1 = Pη−1j=1jyη,η−1,jI . Then, we turn

to the system (Sη,η−1) of dimension (η − 1) × (η − 1) which we integrate with the initial

conditions: θ = 0 and yη,η−1,j(0) = yIη,η−1,j, 1 ≤ j ≤ η − 1. We define θη,η−1 by

θη,η−1 = min{θ > 0 : yη,η−1,η−1(θ) = 0}

and we repeat inductively this procedure for each h ≥ 1, taking as initial conditions for the system (Sη,h) the final conditions for (Sη,h+1).

We define τη = η X h=1 θη,h.

Our aim is to study the asymptotic behaviour of the sequence (τη)η∈IN . We observe first

that, using inductively the approximation theorem of [9], we get immediately the following theorem.

Theorem 1. There exists a function f (., ., .) such that, for any fixed values of the pa-rameters k, η and λ, the size S = S(k, η, λ, n) of the solution found by GRECO when applied with the parameter η to a random instance of m k-sets of size k drawn randomly from a set of size m = nλ−1, satisfies

S

n.f (k, η, λ) → 1 in probability as n → ∞.

We will prove in the appendix that the sequence (τη)η∈IN∗ tends to a limit as η → ∞.

(11)

Theorem 2. There exists a function g(., .) such that, for any fixed values of the pa-rameters k and λ and any positive ǫ, the size S = S(k, η, λ, n) of the solution found by GRECO when applied with the parameter η to a random instance of m k-sets of size k drawn randomly from a set of size m = nλ−1, satisfies

1 − ǫ ≤ S

n.g(k, λ) ≤ 1 + ǫ in probability as n → ∞, if only η is sufficiently large.

3.4 Proof of the convergence of of the sequence (τη)η∈IN

3.4.1 Explicit integration of the differential systems

Let us firs reformulate the system (Sη,h). By introducing β = (k − 1)/k and αη,h =

(1/hk)Rη,h, 1 ≤ h ≤ η, the vectors of IRh : yη,h(θ) = (yη,h,j(θ))T1≤j≤h, eh = (0, 0, . . . , −1)T

and the matrix Mh of dimension h × h with

Mh =      Mh,j,j = −j Mh,j,j+1 = j + 1 Mh,i,j = 0 elsewhere

we can reformulate (Sη,h) in a vectorial form as follows:

(Sη,h) ˙yη,h(θ) = β αη,h− θ Mhyη,h(θ) + eh. We define also τη = η X h=1 θη,h.

Our aim is to study the asymptotic behaviour of the infinite sequence (τη)η∈IN .

We consider (Sη,h) when θ ∈ [0, αη,h[. It is a vectorial first-order linear

non-homogene-ous non-autonomnon-homogene-ous o.d.e. (ordinary differential equation). The system’s coefficients are continuous functions of the variable θ on [0, αη,h[. So, Cauchy’s theorem about the

existence and uniqueness of the solutions of linear Cauchy’s problems is applicable here, and moreover the non-extendable solution of (Sη,h) under the initial conditions yη,h(0) =

yI

η,h, is defined everywhere on [0, αη,h[, ([1], p. IV, 17).

To solve (Sη,h), we consider the associated homogeneous o.d.e.:

(Hη,h) ˙y(θ) =

β αη,h− θ

Mhy(θ).

The resolvant of this homogeneous system is Φ(θ) = exp {β ln (αη,h/αη,h− θ) Mh}. The

calculation of the characteristic polynomial of Mh shows that Mh possesses h distinct

eigenvalues, namely −1, −2, . . . , −h, hence Mh is diagonalizable. If bj is an eigenvector

associated to eigenvalue −j, for each j in {1, 2, . . . , h}, then B = (b1, . . . , bh) is a basis

of IRh and, in terms of B, Mh is expressed by means of the diagonal matrix Dh =



−iδji

(12)

where δji is the Kronecker symbol. Denoting by P the matrix of the change of coordinates

from B to the canonical basis, we have Mh = P DhP−1. For each j, we take bj =

(bj,1, . . . , bj,h)T where bj,i =        (−1)i+jj i  if i < j 1 if i = j 0 if i > j The matrix P is then defined as

Pi,j =        (−1)i+jj i  if i < j 1 if i = j 0 if i > j and its inverse matrix P−1 = Q is defined as

Qi,j =        j i  if i < j 1 if i = j 0 if i > j

Then Φ(θ) = exp {β ln (αη,h/αη,h− θ) P DhP−1} = P exp {β ln (αη,h/αη,h− θ) Dh} P−1,

and we can explicitely compute its components, obtaining so:

     Φi,j(θ) = 0 if i > j Φi,j(θ) = j i   α η,h αη,h−θ −βi 1 − αη,h αη,h−θ −βj−i if i ≤ j (4)

So, we have expicitely integrated (Hη,h). The solution of (Sη,h) under the initial conditions

yη,h(0) = yIη,h is , for θ ∈ [0, αη,h[, ([1], p. IV, 21): yη,h(θ) = Φ(θ)yη,hI + Φ(θ)   θ Z 0 Φ−1(σ)dσ  eh (5) Let uη,h(θ) = Φ(θ)   θ Z 0 Φ−1(σ)dσ  eh. (6)

Since uη,h is the solution of (Sη,h) under the initial conditions uη,h(0) = 0, one can verify

by a straightforward calculation that, for every j = 1, . . . , h and for every θ ∈ [0, αη,h[,

uη,h,j(θ) = − h j ! θ Z 0 αη,h− θ αη,h− t !−βj 1 − αη,h− θ αη,h− t !β  h−j dt. (7)

So, with expressions (4,5,6,7), we have explicitely resolved the system (Sη,h) under the

(13)

3.4.2 The computation of θη,h

Using the results of § 3.4.1, we have, ∀θ ∈ [0, αη,h[,

yη,h,h(θ) = αη,h αη,h− θ !−βh yη,h,hI + 1 βh − 1   α−βh+1η,h (αη,h− θ)−βh − (αη,h− θ)  . Assuming that yI

η,h,h > 0, we search for these θ ∈ [0, αη,h[ verifying yη,h,h(θ) = 0. By a

straightforward computation, we obtain: if yI

η,h,h > 0, k ≥ 2 and h ≥ 2, then there exists

a unique θη,h∈ [0, αη,h[ such that yη,h,h(θη,h) = 0 given by

θη,h= αη,h  1 − 1 + βh − 1 αη,h yη,h,hI ! 1 −βh+1  . (8) 3.4.3 The signs of yI η,h,j

Let us start by establishing the following assertions. (A1) Let 1 ≤ h ≤ η and suppose that ∀j, 1 ≤ j ≤ h, yI

η,h,j > 0. Then:

(i) ∀j ∈ {1, 2, . . . , h − 1}, ∀θ ∈ [0, θη,h[, yη,h,j(θ) > 0;

(ii) ∀j ∈ {1, 2, . . . , h − 1}, yI

η,h−1,j = yη,h,j(θη,h) ≥ 0.

To show (i), we reason by contradiction in assuming that there exist θ ∈ [0, θη,h[ and j,

1 ≤ j ≤ h − 1, such that yη,h,j(θ) ≤ 0. Since yη,h,j(0) = yη,h,jI > 0, we have θ > 0. By an

argument of connectedness, there exists θ1 ∈]0, θ] such that yη,h,j(θ1) = 0. Let us denote

by J = {j : 1 ≤ j ≤ h − 1, yη,h,j([0, θη,h[) ∋ 0}, J 6= ∅. For each j ∈ J, let us introduce

θ(j) = inf{θ ∈ [0, θη,h[ : yη,h,j(θ) = 0} = inf{[0, θη,h] : yη,h,j(θ) = 0}

= inf([0, θη,h] ∩ y−1η,h,j({0})).

Since this last set is closed and bounded below, it contains its greatest lower bound, thus yη,h,j(θ(j)) = 0. As, on the other hand, yη,h,j(0) > 0 and yη,h,j is continuous, then yη,h,j is

positive on a neighbourhood of 0, hence θ(j) = 0. By definition of θ(j), for all θ ∈ [0, θ(j)[, we have yη,h,j(θ) > 0. Let us introduce ˆθ = min{θ(j) : j ∈ J}; we have 0 < ˆθ < θη,h

and let ˆj = max{j ∈ J : θ(j) = ˆθ}. Note that j cannot be equal to h − 1. In fact, if the contrary holds, we have ∀θ ∈ [0, ˆθ[, yη,h,h−1(θ) > 0 and yη,h,h−1(ˆθ) = 0, which implies

˙yη,h,h−1(ˆθ) ≤ 0. By using (Sη,h), we get

0 ≥ ˙yη,h,h−1(ˆθ) = h β αη,h− ˆθ yη,h,h(ˆθ) − (h − 1) β αη,h− ˆθ yη,h,h−1(ˆθ) = h β αη,h− ˆθ yη,h,h(ˆθ) > 0 which is impossible.

Consequently, we have proved that 1 ≤ ˆj ≤ h − 2. Let us note that if ˆj ≤ j ≤ h and if θ ∈ [0, ˆθ], then yη,h,j(θ) > 0 because j > ˆj =⇒ (j /∈ J ∨ θ(j) > ˆθ).

Since yη,h,ˆj(θ) > 0 for θ ∈ [0, ˆθ], and yη,h,ˆj(ˆθ) > 0, we have ˙yη,h,ˆj(ˆθ) ≤ 0. This, by using (Sη,h), implies that

0 ≥ ˙yη,h,ˆj(ˆθ) = (ˆj + 1) β αη,h− ˆθ yη,h,ˆj+1(ˆθ) − ˆj β αη,h− ˆθ yη,h,ˆj(ˆθ) = (ˆj + 1) β αη,h− ˆθ yη,h,ˆj+1(ˆθ) > 0

(14)

a contradiction.

This justifies (i). On the other hand, (ii) is a consequence of (i) because of the continuity of yη,h,j.

(A2) Under the hypotheses of (A1) let 1 ≤ j ≤ h−1. If yη,h,j(θη,h) = 0, then ∀i, j ≤ i ≤ h

we have yη,h,i(θη,h) = 0.

In fact, with (A1) we know that for all θ ∈ [0, θη,h[, yη,h,j(θ) > 0 and yη,h,j(θη,h) = 0,

which implies 0 ≥ ˙yη,h,j(θη,h) = (j + 1) β αη,h− θη,h yη,h,j+1(θη,h) − j β αη,h− ˆθη,h yη,h,j(θη,h) = (j + 1) β αη,h− θη,h yη,h,j+1(θη,h) ≥ 0. Thus, yη,h,j+1(θη,h) = 0.

We can repeat the same reasoning for yη,h,j+2 and so on.

(A3) Under the hypotheses of (A1) we have:

(∀j = 1, . . . , h − 1, yη,h,j(θη,h) > 0) ⇐⇒ (yη,h,h−1(θη,h) > 0).

In fact, the first implication is obvious and the converse results from (A2).

We also remark that yη,h,h−1(θη,h) = 0 ⇐⇒ ˙yη,h,h−1(θη,h) = 0. In fact, it suffices to remark

that by the definition of θη,h, ˙yη,h,h−1(θη,h) = −(h − 1)(β/αη,h− θη,h)yη,h,h−1(θη,h).

3.4.4 Relations between θη,h and αη,h

We multiply the j-th equation of (Sη,h) by j and next, we add the h resulting equations;

we thus obtain Ph

j=1j ˙yη,h,j(θ) = −h − [β/(αη,h− θ)]Phj=1jyη,h,j(θ).

Denoting by zη,h(θ) =Phj=1jyη,h,j(θ), we see that zη,h is a solution of the o.d.e.:

˙zη,h(θ) = −h −

β αη,h− θ

zη,h(θ) (9)

under the initial condition zη,h(0) = Rη,h. We note that equation (9) is an

one-dimen-sional linear non-homogeneous o.d.e. with variable coefficient (continuous function of θ) and consequently, the theorem about the existence and uniqueness of the non-extendable solutions of Cauchy’s problems is usable in (9). Integrating (9), for instance using the method of constants variations, we obtain, ∀θ ∈ [0, αη,h[,

zη,h(θ) = Rη,h 1 − θ αη,h !β + αη, h −β + 1h   1 − θ αη,h ! − 1 − θ αη,h !β = Rη,h− hkθ.

On the other hand, zη,h(θη,h) = Phj=1jyη,h,j(θη,h) = Ph−1j=1 jyIη,h−1,j = Rη,h−1, that implies

Rη,h−1 = Rη,h− khθη,h, 2 ≤ h ≤ n − 1, or equivalently αη,h−1= h h − 1 ! (αη,h− θη,h). (10)

(15)

By introducing the 2-variable function Th(α, y) = α

h

1 − [1 + (βh − 1)/αy]1/(−βh+1)i for α > 0, y > 0, we see, using expression (8), that θη,h= Th(αη,h, yIη,h,h). One can also verify

that if k ≥ 3, h ≥ 2, α > 0, y > 0 and kα ≥ y then 1 −βh + 1ln 1 − −βh + 1 α y ! ≥ 1 1 − (k − 1)(h − 1)ln((k − 1)(h − 1)).

By taking for a short while X = (k − 1)(h − 1) ≥ 2, the previous inequality allows us to obtain [1/(−βh + 1)] ln [1 − (−βh + 1)/αy] ≥ ln(X)/(1 − X) ≥ ln (1/2); therefore, taking the exponential mapping, we get [1 + (βh − 1)/αy][1/(−βh+1)] ≥ 1/2; consequently, if k ≥ 3, h ≥ 2, α > 0, y > 0 and kα ≥ y, then

Th(α, y) ≤

α

2. (11)

From the above expression, since kαη,h = (1/h)Rη,h ≥ yη,h,hI , we deduce, ∀k ≥ 3, ∀h ≥ 2

θη,h≤

1 2αη,h. 3.4.5 An upper bound of yη,h,j

By adding the h equations of the system (Sη,h), we obtain after some calculation the

following expression: Ph

j=1 ˙yη,h,j(θ) = −1 − (β/αη,h)yη,h,1(θ). Since yη,h,1 is non-negative

on [0, θη,h], the derivative of the functionPhj=1yη,h,j is non-positive on the same interval,

hence this function is monotonically decreasing on [0, θη,h], and consequently, ∀θ ∈ [0, θη,h] h X j=1 yη,h,j(θ) ≤ h X j=1 yη,h,jI . (12) By taking θ = θη,h, we getPh−1j=1yη,h−1,jI ≤ Ph

j=1yη,h,jI which implies that ∀h ≤ η h X j=1 yη,h,jI ≤ η X j=1 yη,η,jI ≤ 1. (13) Also, from the above expression and expression (12), we deduce, ∀n ∈ IN∗, ∀h ∈ IN∗, h ≤ η, ∀j ∈ IN∗, j ≤ h, ∀θ ∈ [0, θη,h]

0 ≤ yη,h,j(θ) ≤ 1. (14)

On the other hand, using expressions (5) and (6), we get: yη,h,j(θ) =

h

X

p=1

Φj,p(θ)yη,h,pI + uη,h,p(θ).

We easily verify, using expression (7), that uη,h,p(θ) ≤ 0; thus using expression (4), we

obtain yη,h,j(θ) ≤ h X p=j p j ! αη,h− θ αη,h !jβ 1 − αη,h− θ αη,h !β  p−j yI η,h,p. (15)

(16)

In what follows, we will need inequality (16), whose demonstration is elementary:

xie(1−x)λ≤ 1, x ∈]0, 1], λ > 0, i ∈ IN , i ≥ λ. (16) We are going to establish, by downward induction, inequalities (17) and (18), for all integers η, h, j ∈ IN , λ ≤ j ≤ h ≤ η and ∀θ ∈ [0, θη,h].

yIη,h,j ≤ e

−λλj

j! . (17)

In fact, we initialize the induction at h = η with expression (3). We suppose true the inequalities for a fixed h. Then, in the step h − 1, by putting x = [(αη,h− θη,h)/αη,h]β, we

use expression (15) to obtain, for θ = θη,h,

yη,h−1,jI = yη,h,j(θη,h) ≤ h X p=j p j ! xj(1 − x)p−jyIη,h,p; next, using the induction hypothesis, we get for yI

η,h−1,j: yη,h−1,jI ≤ h X p=j p j ! xj(1 − x)p−je −λλp p! ≤ e−λxjλj j! h X p=j 1 (p − j)!(1 − x) p−jλp−j = e−λλj j! x je(1−x)λ

and, finally, with expression (16) we obtain (17). yη,h,j(θ) ≤

e−λλj

j! . (18)

In fact, by taking now x = [(αη,h− θ)/αη,h]β and by using (15), we get:

yη,h,j(θ) ≤ h X p=j p j ! xj(1 − x)p−jyIη,h,p. Next, using expressions (16) and (17), we obtain

yη,h,j(θ) ≤ h X p=j p j ! xj(1 − x)p−je −λλp p! ≤ e−λλj j! x je(1−x)λ e−λλj j! q.e.d.

Let us now introduce the constant

Ξ = 1 min j∈IN,1≤j≤λ n e−λλj j! o. (19)

We note that Ξ does not depend on λ and also that Ξ > 1, since e−λλj/j! < e−λeλ = 1.

Using Ξ, we can prove the following expression. 0 ≤ yη,h,j(θ) ≤ Ξ

e−λλj

(17)

In fact, if j ≥ λ, then by using (18), (Ξ > 1) we get yη,h,j(θ) ≤ e−λλj/j! ≤ Ξe−λλj/j!. On

the other hand, if j < λ, then by expression (14) we get yη,h,j(θ) ≤ 1 ≤ Ξ

e−λλj

j! q.e.d.

3.4.6 An upper bound of the sequence (τη)η∈IN

Using expression (10) and for h ≥ 2, we get: θη,h= αη,h− h − 1 h αη,h−1 = αη,h− αη,h−1+ 1 hαη,h−1. Consequently, τη = η X h=1 θη,h= θη,1+ η X h=2 (αη,h− αη,h−1) + η X h=2 1 hαη,h−1 = θη,1+ αη,η− αη,1+ η−1 X h=1 1 h + 1αη,h. Thus, τη = (θη,1− αη,1) + αη,η+ η−1 X h=1 1 h + 1αη,h. We know that θη,1 ≤ αη,1 and, by using expression (3) we obtain

τη ≤ 1 ηk η X j=1 je −λλj j! + η−1 X h=1 1 h + 1 1 hk h X j=1 jyη,h,jI ≤ e −λ η λ k η−1 X j=0 λj j! + η−1 X h=1 1 (h + 1)hk h X j=1 jΞe −λλj j! the last inequality obtained using (20); hence

τη ≤ e−λ η λ k η−1 X j=0 λj j! + Ξe−λλ k η−1 X h=1 1 h(h + 1) h−1 X j=0 λj j!. Note that η−1 X h=1 1 h(h + 1) h−1 X j=0 λj j! ≤ e λ η−1 X h=1 1 h(h + 1) = e λ η−1 X h=1 1 h − 1 h + 1  = eλ 1 − 1 η ! which leads to τη ≤ e−λλ k   1 η η−1 X j=0 λj j!  + Ξλ k 1 − 1 η ! . (21)

Note also that (1/η)Pη−1

j=0(λj/j!) ≤ max0≤j≤n−1{λj/j!}. By denoting by ⌊λ⌋ the floor of λ,

we easily verify that: max0≤j≤λ−1{λj/j!} = λ⌊λ⌋−1/(⌊λ⌋ − 1)! and maxλ−1≤j{λj/j!} =

λ⌊λ⌋/⌊λ⌋! and, since λ ≥ ⌊λ⌋, we have λ⌊λ⌋/⌊λ⌋! ≥ λ⌊λ⌋−1/(⌊λ⌋ − 1)!; consequently, we

get maxj∈IN {λj/j!} = λ⌊λ⌋/⌊λ⌋!. Whenever 0 < λ < 1, we have ⌊λ⌋ = 0, ⌊λ⌋! = 1 and

λj/j! ≤ 1 = λ0/0! = λ⌊λ⌋/⌊λ⌋!, therefore, ∀λ > 0, max j∈IN ( λj j! ) = λ ⌊λ⌋ ⌊λ⌋!. (22)

(18)

On the other hand, 1 − (1/η) ≤ 1 and by (21) and (22) we deduce that τη ≤ e−λλ k λ⌊λ⌋ ⌊λ⌋! + Ξλ k = λ k " e−λλ⌊λ⌋ ⌊λ⌋! + Ξ # and consequently, sup η≥2τη ≤ λ k " e−λλ⌊λ⌋ ⌊λ⌋! + Ξ # < +∞. (23)

Since Ξ depends only on λ (see also (19)), the proposed upper bound only depends on λ and k.

3.4.7 Variations with respect to α

We consider for each h ∈ IN∗ and for each α > 0, the following differential system in IRh: (Ch,α) ˙y(θ) =

β

α − θMhy(θ) + eh. when θ ∈ [0, α[.

One can see that (Sη,h) = (Ch,αη,h).

We denote by ϕh(θ, α, zI) the value at the time θ of the solution of (Ch,α) that takes the

value zI at the time 0. Then, function T

h(α, zhI) introduced above denotes the first time

the h-th coordinate (ϕh,h) of ϕh becomes 0. We can see that yη,h(θ) = ϕ(θ, αη,h, yη,hI ),

∀θ ∈ [0, αη,h[. We have already noted that θη,h= Th(αη,h, yη,h,hI ).

The homogeneous system associated with (Ch,α) is denoted (Lh,α) and the resolvant

of (Lh,α) is denoted by Φh(θ, α).

With computations similar to the ones performed for the systems (Sη,h), we establish the

following: Φh(θ, α) = exp {β ln [α/(θ − α)] Mh}, with

Φh,i,j(θ, α) =      0 if i > j  j i   α α−θ −βi 1 −α−θα −β j−i if i ≤ j (24)

We have also ∀j = 1, . . . , h, ϕh,j(θ) ≥ 0 in [0, Th(α, zhI)] when zIh,i > 0.

The solution of (Ch,α) that is equal to zIat the time 0 is ϕ(θ, α, zI) = Φh(θ, α)zI+ψh(θ, α),

where ψh,j(θ, α) = − h j ! Z θ 0 α − θ α − t !βj 1 − α − θ α − t !β  h−j dt. (25)

Starting from expression (25), we have, ∀j ≤ h and ∀θ ∈ [0, Th(α, zhI)]

ψh,j(θ) ≤ 0. (26)

Moreover, with a similar reasoning as in the case of expression (18), if ∀j = 1, . . . , h, zI

j ≤ Ξe−λλj/j!, then ∀j = 1, . . . , h, ∀θ ∈ [0, Th(α, zhI)] we have

ϕh,j(θ, α, zI) ≤ Ξ

e−λλj

(19)

Also by arguments similar to the ones of § 3.4.4 we establish that function θ 7→ h X j=1 jϕh,j(θ, α, zI)

is solution of the scalar o.d.e. ˙x(θ) = −h − [β/(α − θ)]x(θ). We then deduce that

h X j=1 jϕh,j(θ, α, zI) =   h X j=1 jzjI   1 − θ α !β + hα −β + 1   1 − θ α ! − 1 − θ α !β .

Supposing also that Ph

j=1jzjI ≤ hkα, after some easy algebra we obtain: ∀h ∈ IN∗,

∀α > 0, ∀zI ∈ IR+h such that Ph j=1jzIj ≤ hkα h X j=1 jϕh,j(θ, α, zI) ≤ hk(α − θ). (28)

We now compute a lower bound for Rη,η and αη,η. In fact, we prove that ∀λ > 0,

∃η(λ) ∈ IN∗ such that ∀η ∈ IN∗, η ≥ η(λ) Rη,η ≥ λ 2 αη,η ≥ λ 2kη. (29)

Since limη→∞Pη−1i=0(λi/i!) = eλ > eλ/2 and

 Pη−1

i=0(λi/i!)



η is a monotonic increasing

sequence, then ∃n(λ) ∈ IN∗ such that ∀η ≥ η(λ), Pη−1

i=0(λi/i!) ≥ eλ/2. Consequently,

Rη,η = e−λλ η−1 X i=0 λi i! ≥ e −λλeλ 2 = λ 2 and αη,η = Rη,η ηk ≥ λ 2kη q.e.d.

Let us now prove the following: ∀η ∈ IN∗, ∀h ∈ IN∗, h ≤ η

yη,h,hI ≥ θη,h. (30)

Recall that

0 ≤ ∂Th

∂y (α, y) ≤ 1 (31)

and also that ∀α > 0, Th(α, 0) = 0. Using the mean value theorem, we have

0 ≤ θη,h = Th(αη,h, yη,h,hI ) = |Th(αη,h, yη,h,hI ) − Th(αη,h, 0)|   sup y∈[0,yI η,h,h] ∂Th ∂y (αη,h, y)  |yIη,h,h− 0| ≤ 1|yIη,h,h| = yη,h,hI .

(20)

This concludes the proof of expression (30).

Moreover, by some easy calculations, we can prove the following ∀p ∈ IN∗, p ≥ 2λ, ∀h, η ∈ IN∗ such that η − 1 ≥ h ≥ 2λ λp p! ≤ λp−1 (p − 1)! − λp p! η−1 X p=h λp p! ≤ λh (h − 1)! − λη−1 (η − 1)!. (32)

We now prove that ∀η ∈ IN∗, ∀h ∈ IN∗ such that h < η we have

hαη,h≥ (h + 1)αη,h+1− (h + 1)yη,h+1,h+1I . (33)

In fact, using expression (10), we have Rη,h= Rη,h+1− (h + 1)kθη,h+1 or

Rη,h

k =

Rη,h+1

k − (h + 1)θη,h+1. Recall also that αη,h = Rη,h/(hk). So, we get

hαη,h= (h + 1)αη,h+1− (h + 1)θη,h+1.

Moreover, using expression (30): −θη,h+1 ≥ −yη,h+1,h+1I . Combining the last two

expres-sions, we obtain expression (33).

Let us now prove that ∀η ∈ IN∗, η ≥ 2λ and ∀h ∈ IN∗, 2λ ≤ h ≤ η we have αη,h≥

η hαη,η−

e−λλh

h! . (34)

Using expression (33), ∀l ∈ IN∗ such that η > l ≥ h ≥ 2λ, we have lαη,l− (l + 1)αη,l+1 ≥ −(l + 1)yIη,l+1,l+1 ≥ −(l + 1) e−λλl+1 (l + 1)! (17) ≥ e −λλl+1 l! = e −λλλl l!. Summing these inequalities, we obtain (using expression (32))

η−1 X l=h lαη,l− η−1 X l=h (l + 1)αη,l+1≥ −e−λλ η−1 X l=h λl l!

Using also expression (32), we get hαη,h− ηαη,η ≥ −e−λλh/(h − 1)!, and expression (34)

follows.

Also, ∀η ∈ IN∗, η > 2λ, η > η(λ) (expression (29)) and ∀h ∈ IN∗ such that 2λ ≤ h ≤ η, we have αη,h≥ λ 2hk − e−λλh h! . (35)

(21)

In fact, using expressions (34) and (29), αη,h ≥ (η/h)αη,η− e−λλh/h! ≥ (η/h)(λ/2ηk) −

e−λλh/h! and expression (35) follows.

We finally prove that ∃h∗ ∈ IN∗ such that ∀η ∈ IN∗, η > 2λ, η ≥ η(λ) (expression (29)),

∀η ∈ IN∗, η > h ≥ h∗ , h ≥ 2λ, we have

αη,h≥

λ

4hk. (36)

Since λh−1/(h − 1)!

h is the general term of a convergent series, λ

h−1/(h − 1)! →

h→∞ 0.

Thus, λh/(h − 1)! = λλh−1/(h − 1)! →

h→∞ λ.0 = 0 or hλh/h! = λh/(h − 1)! →h→∞ 0;

consequently, ∃h∗ ∈ IN∗ such that ∀h ≥ h∗, hλh/h! ≤ λ/(4ke−λ) or e−λλh/h! ≤ λ/(4hk).

Thus, using expression (35), we deduce expression (36).

Let us now introduce, ∀h ∈ IN∗, the norm k · kh in IRh, defined as kxkh =Phi=1(i/h)|xi|.

Then, ∃c1 > 0 such that ∀h ∈ IN∗, h > 2λ, ∀η ∈ IN∗, η ≥ h, ∀α ≥ min{αη,h, αη+1,h},

∀zI ∈ IR+h such that ∀j = 1, . . . , h, zI

j ≤ Ξe−λλj/j! and zhI ≤ kα we have

∂ ∂θϕh(Th(α, z I h), α, zI) h ≤ c1. (37)

Let us fix η and h such that η ≥ h > 2λ. Let also θ = Th(α, zhI).

Then, ∀j = 1, . . . , h − 1, we have j h ∂ ∂θϕh,j(θ, α, z I) ≤ j(j + 1) h β α − θ(ϕh,j+1(θ, α, z I) + ϕ h,j(θ, α, zI)). Since zI

h ≤ kα (expression (11), we have θ ≤ (1/2)α, hence β/(α − θ) ≤ 2β/α.

Let us suppose that α ≥ αη,h, the reasoning being similar in the case α ≥ αη+1,h. So,

β/(α − θ) ≤ 2β/αη,h and since h > 2λ, using expression (36), we get β/(α − θ) ≤

8βhk/λ = (8(k − 1)/λ)h. Thus, using the hypotheses of expression (37) and expres-sion (27), we have j h ∂ ∂θϕh,j(θ, α, z I) ≤ j(j + 1) h 8(k − 1) λ h(ϕh,j+1(θ, α, z I) + ϕ h,j(θ, α, zI)) ≤ 8(k − 1) λ j(j + 1) Ξ e−λλj+1 (j + 1)! + Ξ e−λλj j! ! . Hence, h X j=1 j h ∂ ∂θϕh,j(θ, α, z I) ≤ 8(k − 1) λ Ξe −λh−1X j=1 j(j + 1) λ j j! + λj+1 (j + 1)! ! ≤ 8(k − 1) λ Ξe −λ2eλ2+ λ) ≤ 16(k − 1)Ξ(λ + 1).

On the other hand, h h| ∂ ∂θϕh,h(θ, α, z I)| ≤ 1 + β α − θϕh,h(θ, α, z I) ≤ 1 +8(k − 1) λ hkΞ e−λλh h! ≤ 1 + 8(k − 1)Ξ

(22)

and, finally, h X j=1 j h ∂ ∂θϕh,j(θ, α, z I) ≤ 8(k − 1)Ξ[2(λ + 1) + 1] + 1 = c1 q.e.d.

We now prove that ∃c2 > 0 such that ∀h ∈ IN∗, h ≤ 2λ, ∀η ∈ IN∗, η ≥ h, ∀α > 0,

∀zI ∈ IR+h such that Ph j=1jzjI ≤ hkα, we have ∂ ∂θϕh(Th(α, z I h), α, zI) h ≤ c2. (38)

Let us fix η, h, α and put θ = Th(α, zhI). If j ≤ h − 1, then

∂ ∂θϕh,j(θ, α, z I) = β α − θ h (j + 1)ϕh,j+1(θ, α, zI) − jϕh,j(θ, α, zI) i ≤ β α − θ[(j + 1)ϕh,j+1(θ, α, z I) + jϕ h,j(θ, α, zI)] ≤ β α − θ h X i=1 iϕh,i(θ, α, zI) and using (28), β α − θ h X i=1 iϕh,i(θ, α, zI) ≤ β α − θhk(α − θ) = βhk = h(k − 1) ≤ 2λ(k − 1) ∂ ∂θϕh,h(θ, α, z I) ≤ 1 + β α − θϕh,h(θ, α, z I) ≤ 1 + β α − θhk(α − θ) = 1 + h(k − 1) ≤ 2λ(k − 1) + 1. So, ∂ ∂θϕh(θ, α, z I) h = h X i=1 i h ∂ ∂θϕh,i(θ, α, z I) ≤ h X i=1 h h ∂ ∂θϕh,i(θ, α, z I) = h−1 X i=1 ∂ ∂θϕh,i(θ, α, z I) + ∂ ∂θϕh,h(θ, α, z I) ≤ h−1 X i=1 2λ(k − 1) + 2λ(k − 1) + 1 ≤ (2λ)2(k − 1) + 2λ(k − 1) + 1 = c2 q.e.d.

Using (37) and (38), putting c3 = max{c1, c2}, noting that hzhI ≤

Ph

i=1jzjI ≤ hkα, or

zI

h ≤ kα, we deduce that ∃c3, ∀h ∈ IN∗, ∀η ∈ IN∗, η ≥ h, ∀α ≥ min{αη,h, αη+1,h},

∀zI ∈ IR+h such that ∀j = 1, . . . , h, zI jΞe−λλj/j! and Ph i=1jzjI ≤ hkα we have ∂ ∂θϕh(Th(α, z I h), zI) h ≤ c3. (39)

(23)

We now prove that ∀v ∈ IRh, ∀α > 0, ∀θ ∈ [0, Th(α, v)] kΦh(θ, α)vkh ≤ kvkh. (40) We have kΦh(θ, α)vkh = h X i=1 i h h X j=1 Φh,i,j(θ, α)vj ≤ h X i=1 i h h X j=1 |Φh,i,j(θ, α)| |vj| .

By expression (24), putting x = [(α − θ)/α]β (0 ≤ x ≤ 1 or 1 − x ≥ 0), we have Φh,i,j(θ, α) ≥ 0. So, kΦh(θ, α)vkh ≤ h X i=1 i h h X j=1 Φh,i,j(θ, α) |vj| = 1 h h X j=1 h X i=1 i j! i!(j − i)!x i(1 − x)j−i|v j| = i h h X j=1 jx   j−1 X l=0 j − 1 l ! xl(1 − x)j−1−l  |vj| = 1 h h X j=1 jx(x + 1 − x)j−1|vj| = xkvkh q.e.d.

Next, we prove that ∀h ∈ IN∗, ∀α > 0, ∀θ < α

1 αψh(θ, α) h ≤ 1. (41)

Let us put x(t) = [(α − θ)/(α − t)]β (x(t) ≤ 1). Then

1 αψh(θ, α) h = 1 α h X i=1 i h|ψh,i(θ, α)| = 1 α h X i=1 i h θ Z 0 x(t)i(1 − x(t))h−idt = 1 α θ Z 0 " h X i=1 h − 1 i − 1 ! x(t)i−1(1 − x(t))h−1−(i−1)x(t) # dt = 1 α θ Z 0 (x(t) + 1 − x(t))h−1x(t)dt = 1 α θ Z 0 x(t)dt = 1 α θ Z 0 α − θ α − t !β dt ≤ θ α ≤ 1 q.e.d. Also, ∀h ∈ IN∗, ∀α > 0, ∀zI ∈ IRh, ∀θ ∈ [0, T h(α, zhI)], we have ∂ ∂αϕh(θ, α, z I) = −θ α ∂ ∂θϕh(θ, α, z I) + 1 αψh(θ, α). (42)

We fix h, α, zI. We have already seen that θ ∈ [0, T

h(α, zhI)] and also that

(24)

Let us put χh(θ, α, zI) = Φh(θ, α)zI, so,

ϕh(θ, α, zI) = χh(θ, α, zI) + ψh(θ, α).

Let us introduce the function w : [0, 1[7→ IRh, w(s) = exp {β ln [1/(1 − s)] Mh} zI; then,

χh(θ, α, zI) = w (θ/α).

We also introduce the function v : [0, 1[7→ IRh, v(s) = (v1(s), . . . , vh(s))T, where

vj(s) = − h j ! s Z 0 1 − s 1 − r βj" 1 − 1 − s 1 − r β#h−j dr.

Performing a variable changing t = αr, we obtain ψh(θ, α) = αv (θ/α).

From the above, we deduce that ϕh(θ, α, zI) = w (θ/α) + αv (θ/α).

Therefore, ∂ ∂θϕh(θ, α, z I) = ˙w θ α ! 1 α + α ˙v θ α ! 1 α = ˙w θ α ! 1 α + ˙v θ α ! . On the other hand,

∂ ∂αϕh(θ, α, z I) = ˙w θ α ! −θ α2 ! + v θ α ! + ˙v θ α ! −θ α2 !

and after some easy algebra, expression (42) is proved.

We now prove the following: ∃c4 > 0 such that ∀h ∈ IN∗, ∀η ∈ IN∗, η ≥ h, we have

kϕh+1(θη+1,h+1, αη+1,h+1, yIη+1,h+1) − ϕh+1(θη,h+1, αη,h+1, yIη+1,h+1)kh+1

≤ c4|αη+1,h+1− αη,h+1|. (43)

To shorten the writing, we put zI = yI

η+1,h+1. So, θη+1,h+1 = Th+1(αη+1,h+1, zh+1I ) and

θη,h+1 = Th+1(αη,h+1, zh+1I ). Let us also denote by [x1; x2] a convex segment (notation

which does not signify that x1 ≤ x2).

Let us introduce, for α ∈ [αη,h+1; αη+1,h+1], the function g(α) = ϕh+1(Th+1(α, zh+1I ), α, zI).

Thus

kϕh+1(θη+1,h+1, αη+1,h+1, yIη+1,h+1) − ϕh+1(θη,h+1, αη,h+1, yIη+1,h+1)kh+1

= kg(αη+1,h+1) − g(αη,h+1)kh+1.

Using the mean value theorem, we get

kg(αη+1,h+1) − g(αη,h+1)kh+1≤ sup α∈[αη,h+1;αη+1,h+1]

k ˙g(α)kh+1

!

|αη+1,h+1− αη,h+1|.

On the other hand ˙g(α) = ∂ϕh+1 ∂θ (Th+1(α, z I h+1), α, zI) ∂Th+1 ∂α (α, z I h+1) + ∂ϕh+1 ∂α (Th+1(α, z I h+1), α, zI) (42) = ∂Th+1 ∂α (α, z I h+1) − θ α ! ∂ϕh+1 ∂θ (Th+1(α, z I h+1), α, zI) + 1 αψh+1(Th+1(α, z I h+1), α).

(25)

Consequently k ˙g(α)kh+1 ≤ ∂Th+1 ∂α (α, z I h+1) + θ α ! ∂ϕh+1 ∂θ (Th+1(α, z I h+1), α, zI) h+1 + 1 αψh+1(Th+1(α, z I h+1), α) h+1 .

Let us remark here that one can prove by elementary techniques that

∂Th(α, y) ∂α ≤ 4. (44)

So, using this inequality as well as θ < α and (41), we get k ˙g(α)kh+1≤ 5 ∂ϕh+1 ∂θ (Th+1(α, z I h+1), α, zI) h+1 + 1.

We can also remark that the hypotheses made on α and zI for the proof of (39) remain

valid here also; therefore, ∀α ∈ [αη,h+1; αη+1,h+1], we have k ˙g(α)kh+1≤ 5c3+ 1, q.e.d.

The following inequality also holds ∀η, h ∈ IN∗, h ≤ η: |αη+1,h− αη,h| ≤ 1 kky I η+1,h− yIη,hkh. (45) In fact, |αη+1,h− αη,h| ≤ 1 hk h X i=1 iyIη+1,h,i− 1 hk h X i=1 iyIη+1,h,i = 1 k h X i=1 i hy I η+1,h,i− h X i=1 i hy I η,h,i = 1 k|ky I η+1,hkh− kyη,hI kh| ≤ 1 kky I η+1,h− yη,hI kh

the last two inequalities holding because yI

η+1,h, yη,hI ∈ IRh and because k · kh is a norm

of IRh respectively, q.e.d.

We now prove that ∀η, h ∈ IN∗, h < η, we have

kϕh+1(θη+1,h+1, αη+1,h+1, yIη+1,h+1) − ϕh+1(θη,h+1, αη,h+1, yIη+1,h+1)kh+1

≤ kyIη+1,h+1− yη,h+1I kh+1. (46)

Plainly

kϕh+1(θη+1,h+1, αη+1,h+1, yIη+1,h+1) − ϕh+1(θη,h+1, αη,h+1, yIη+1,h+1)kh+1

≤ kΦh+1(θη,h+1, αη,h+1)(yIη+1,h+1− yη,h+1I )kh+1.

Using this inequality and expression (40), we obtain (46).

Let us continue by showing the following: ∃c5 > 0 such that ∀η, h ∈ IN∗, h < η, we have

(26)

Let us start the proof of expression (47) by noting that, if x = (x1, . . . , xh, 0)T is a vector

of IRh+1 and if we denote by x′ the sub-vector (x

1, . . . , xh)T of x, then kx′kh =  1 + 1 h  kxkh+1≤ 2kxkh+1. kyI η+1,h− yη,hI kh = kyη+1,h+1(θη+1,h+1) − yη,h+1(θη,h+1)kh ≤ 2kϕh+1(θη+1,h+1, αη+1,h+1, yη+1,h+1I ) − ϕh+1(θη,h+1, αη,h+1, yη,h+1I )kh+1≤ 2kϕh+1(θη+1,h+1, αη+1,h+1, yη+1,h+1I ) − ϕh+1(θη,h+1, αη,h+1, yη+1,h+1I )kh+1 +2kϕh+1(θη,h+1, αη,h+1, yη+1,h+1I ) − ϕh+1(θη,h+1, αη,h+1, yη,h+1I )kh+1 (26,45,46) ≤ 2c 4 k + 1  kyI η+1,h+1− yη,h+1I kh+1 = c5kyη+1,h+1I − yη,h+1I kh+1 q.e.d.

Also, ∃c6 > 0 such that ∀η ∈ IN∗, η > 2λ, we have

sup θ∈[0,θη+1,η+1] k ˙yη+1,η+1(θ)kη+1≤ c6. (48) Let j = 1, . . . , η; then ˙yη+1,η+1,j(θ) = (j + 1) β αη+1,η+1− θ yη+1,η+1,j+1(θ) −j β αη+1,η+1− θ yη+1,η+1,j(θ) or j η + 1| ˙yη+1,η+1,j(θ)| ≤ j(j + 1) η + 1 β αη+1,η+1− θ (yη+1,η+1,j(θ) + (yη+1,η+1,j+1(θ)). Also, β αη+1,η+1− θ ≤ β αη+1,η+1− θη+1,η+1 ≤ 2β αη+1,η+1 ≤ 8β(n + 1)k λ = 8(n + 1)(k − 1) λ .

Therefore, using (20), we get j η + 1| ˙yη+1,η+1,j(θ)| ≤ 8(k − 1) λ Ξe −λj(j + 1) " λj j! + λj+1 (j + 1)! # or, η X j=1 j η + 1| ˙yη+1,η+1,j(θ)| ≤ 8(k − 1) λ Ξe −λ2eλ2+ λ) = 16(k − 1)Ξ(λ + 1).

On the other hand, ˙yη+1,η+1,η+1(θ) ≤ 1 + β αη+1,η+1− θ yη+1,η+1,η+1(θ) ≤ 1 + 8(k − 1)Ξ e−λλη η! ≤ 1 + 8(k − 1)Ξ.

(27)

Thus finally, k ˙yη+1,η+1(θ)kη+1 ≤ 8(k − 1)Ξ(4(λ + 1) + 1) + 1 = c6, q.e.d.

Let us now prove that ∃c7 > 0 such that ∀η ∈ IN∗, η ≤ 2λ − 1, we have

sup θ∈[0,θη+1,η+1] k ˙yη+1,η+1(θ)kη+1≤ c7. (49) Let j = 1, . . . , η; then ˙yη+1,η+1,i(θ) ≤ β αη+1,η+1− θ

((i + 1)yη+1,η+1,i+1(θ) + i

β αη+1,η+1− θ yη+1,η+1,i(θ)) ≤ β αη+1,η+1− θ h X j=1 jyη+1,η+1,j (13) ≤ β αη+1,η+1− θ (h + 1)k(αη+1,η+1− θ) ≤ β(n + 1)k < 2λ(k − 1). On the other hand

˙yη+1,η+1,η+1(θ) ≤ 1 + β αη+1,η+1− θ yη+1,η+1,η+1(θ) ≤ 1 + β αη+1,η+1− θ η+1 X j=1 jyη+1,η+1,j(θ) (13) = 1 + β αη+1,η+1− θ (n + 1)k(αη+1,η+1− θ) = 1 + β(n + 1)k ≤ 1 + (k − 1)2λ.

Hence, k ˙yη+1,η+1(θ)kη+1 ≤ [Pηi=1(i/η + 1)][(k − 1)2λ]+ [1 + (k − 1)2λ] ≤ 4λ(k − 1) + 1 = c7,

q.e.d.

Using (48) and (49) and putting c8 = max{c6, c7}, we obtain the following, ∀η ∈ IN∗

sup

θ∈[0,θη+1,η+1]

k ˙yη+1,η+1(θ)kη+1≤ c8. (50)

We now prove that ∃c9 > 0 such that ∀η ∈ IN∗, we get

kyI

η+1,η− yη,ηI kη ≤ c9

e−λλη+1

(η + 1)!. (51)

Using the remark in the beginning of the proof of (47), we have kyI η+1,η− yIη,ηkη = kyη+1,η+1(θη+1,η+1) − (yη,ηI , 0)kη ≤ 2kyη+1,η+1(θη+1,η+1) − (yIη,η, 0)kη+1 = 2 η X i=1 i η + 1 yη+1,η+1,i(θη+1,η+1) − e−λλi i! ≤ 2 η+1 X i=1 i η + 1 yη+1,η+1,i(θη+1,η+1) − e−λλi i!

(28)

= 2 η+1 X i=1 i η + 1|yη+1,η+1,i(θη+1,η+1) − yη+1,η+1,i(0)| = 2 η+1 X i=1 kyη+1,η+1(θη+1,η+1) − yη+1,η+1(0)kη+1 = 2 θη+1,η+1 Z 0 ˙yη+1,η+1(θ)dθ η+1 ≤ 2 θη+1,η+1 Z 0 k ˙yη+1,η+1(θ)kη+1dθ (50) ≤ 2c8 θη+1,η+1 Z 0 dθ = 2c8θη+1,η+1 ≤ 2c8 e−λλη+1 (η + 1)! (30) = c9 e−λλη+1 (η + 1)! q.e.d.

Finally, let us prove the following: ∀η, h ∈ IN∗, h ≤ η, we have |θη+1,h− θη,h| ≤ 4  1 + 1 k  kyη+1,hI − yη,hI kh. (52) By (44) and (31), we have |θη+1,h− θη,h| = |Th(αη+1,h, yη+1,h,hI ) − Th(αη,h, yη,h,hI )| ≤ 4(|αη+1,h− αη,h| + |yIη+1,h,h− yIη,h,h|).

So, using (45), we obtain expression (52).

3.4.8 The convergence of the sequence (τη)η∈IN

We prove first that the series of general term (Pη

h=1|θη+1,h− θη,h|)η≥1is convergent in IR+.

In fact, this series is a positive terms one, hence in order to prove its convergence we shall use the comparison principle.

By combining expressions (52), (51) and (47), we can obtain

η X h=1 |θη+1,h− θη,h| ≤ η X h=1 cη−h5 c9 e−λλη+1 (η + 1)! = 4  1 + 1 k  c9cη5 e−λλη+1 (η + 1)! η X h=1 1 c5 h . Since c5 ≥ 2 > 1, we get η X h=1 |θη+1,h− θη,h| ≤ 4  1 + 1 k c 9 c5 e−λ(λc 5)η+1 (η + 1)! 1 c5− 1 .

Putting c10 = 4 [1 + (1/k)] (c9/c5)[1/(c5− 1)] > 0, the quantity Pηh=1|θη+1,h − θη,h| is

bounded above by c10e−λ(λc5)η+1/(η + 1)!; hence, ∞ X η=1 η X h=1 |θη+1,h− θη,h| ! ≤ c10 e−λ(λc 5)η+1 (η + 1)! ≤ c10e −λeλc5 < +∞

(29)

q.e.d.

Let us now prove that the series of general term (θη+1,η+1)η≥1 converges in IR+.

Plainly, using (30), we get θη+1,η+1≤ yIη+1,η+1,η+1 = e−λ(λ)η+1/(η + 1)!.

Since it is a positive term series, it suffices to apply the comparison principle, by noting that the series of general term (e−λλη+1/(η + 1)!)

η≥1 converges in IR+, q.e.d.

The series of general term (|τη+1− τη|)η≥1 converges in IR+.

In fact, |τη+1− τη| = η+1 X h=1 θη+1,h− η X h=1 θη,h = θη+1,η+1+ η X h=1 (θη+1,h− θη,h) ≤ θη+1,η+1+ η X h=1 |θη+1,h− θη,h|.

The convergence follows from the comparison principle and the propositions proved above. We conclude the section by proving that the sequence (τη)η∈IN∗ converges in IR+.

Using the proposition proved above, the series of general term (τη+1− τη)η≥1 is absolutely

convergent in IR, thus convergent in IR.

Let us remark that τη =Pη−1q=1(τq+1− τq) + τ1. Moreover,

lim η→∞τη = ∞ X q=1 (τq+1− τq) + τ1.

These two remarks conclude the convergence of the sequence (τη)η∈IN∗ in IR+.

4

Set packing and hypergraph independent set

4.1 The PACK algorithm

We consider the greedy algorithm 3 for finding a set packing for P. Informally, PACK algorithm selects at each step one set chosen at random between the sets of smallest cardinality which do not intersect any previously selected set. By exchanging the two color classes of the incidence graph, this algorithm gives an independent set in the hypergraph P, Q, i.e. a set of vertices no two of which belong to the same hyperedge.

4.2 Analysis of PACK on random instances 4.2.1 Analysis of PACK on standard distribution

The analysis of PACK on random instances can be carried over along the same lines than for GRECO. However, this leads prima facie to a more complicated system of differential equations because, in the case of PACK, we have three succesive removals at each step (see the description of PACK).

(30)

[Step1] choose at random one of the Xi’s with minimum di;

add Xi to the solution, delete it and delete also all the classes in Q, say Yi1, . . . , Yih,

incident to it in the graph Π;

next delete any other vertex in X adjacent to one of the classes Yi1, . . . , Yih;

let X′, Y, P, Qdenote the analogues of X, Y, P, Q for the new configuration;

let Π′ denote the subgraph of Π induced by X′∪ Y′;

[Step2] if X′ is empty, then stop, else, execute Step1 with X, Y, P, Q, Πin place of

X, Y, P, Q, Π.

Algorithm 3. PACK algorithm.

of PACK, the conditional distribution of the remaining sets X′

1, . . . , Xh′ (relatively to the

previous unraveling of the algorithm) coincides with the uniform distribution; we can then compute the expectation of the number of these sets which will be tried before finding one which does not intersect the partial packing as a function t(j) of the size j of the partial packing which satisfies t(j) ∼ [1 − (jk/n)]−k.

Thus, the expectation of the number T (s) of trials before a packing of size s is obtained satisfies ET (s) = s X j=1 t(j) ∼ n k(k − 1). " 1 (1 −skn)k−1 − 1 # . (53)

Using the fact that the trials are independent, Tchebichev’s inequality implies T (s) ∼ ET (s) and solving (53) for s, one gets the following estimate for the size of the obtained packing: s = n k  1 − 1 [1 + αk(k − 1)]k−11  .

4.2.2 The differential system simulating PACK for solving the hypergraph independent set problem

When analysing PACK on the hypergraph independent set problem, we interchange the color classes of the incidence graph and, since it is not true that for any given value h the joint distribution of the sets Yi with |Yi| = h is uniform, we must abandon the

preceding stopping times method and come back to Markov chains. The expectations of the increments during the stage when vertices of degree h enter the independent set are

E∆mj = ((j + 1)mj+1− jmj)S, j 6= h,

E∆mh = −1 + ((h + 1)mh+1− hmh))Sd,

where S = h(k − 1)/(P

imi).

With the same notation as in § 3.3, we get the differential system ˙yn,h,j = (j + 1)yn,h,j+1 h(k − 1) Rn,h− hkθ − jyn,h,j h(k − 1) Rn,h− hkθ , 0 ≤ j ≤ h − 1,

(31)

˙yn,h,h = −1 + (h + 1)yn,h,h+1 h(k − 1) Rn,h− hkθ − hyn,h,h h(k − 1) Rn,h− hkθ .

A similar method, as for system (Snh) in the case of hitting set, could be developed to

solve the above system.

5

Conclusions

We have proposed a general method to analyse the behaviour of approximation algorithms for hard optimization problems defined on random set systems. We note that this method is quite general and its application could be extended even to graph problems. In fact, a graph G = (V, E) can be seen as a hypergraph whose vertex set is E, each vertex of V being a hyperedge containing its incident edges. Hence, problems like minimum dominating set, minimum vertex covering, maximum independent set, etc., could be solved by algorithms, the behaviour of which could be analyzed using the framework described.

We remark here that, when working on average case approximation, a well-known difficult mathematical problem is the precise estimation of the expectation of the size of the optimal solution. Such an estimation would enrich our method, by permitting to obtain close estimations for the expected values of the approximation ratios for the algorithms under study.

References

[1] N. Bourbaki, Fonctions d’une Variable R´eelle (Hermann, Paris, 1976).

[2] W. Fernandez de la Vega, Sur la cardinalit´e maximum des couplages d’hypergraphes al´eatoires uniformes, Discrete Mathematics 40, (1982) 315-318.

[3] W. Fernandez de la Vega, Induced trees in sparse random graphs, Graphs and Com-binatorics 2, (1986) 227-231.

[4] W. Fernandez de la Vega, V. Th. Paschos, R. Saad, Average case analysis of a greedy algorithm for the minimum hitting set problem, in: Proc. LATIN’92, Lecture Notes in Computer Science, Vol. 583 (Springer, Berlin, 1992) 130-138.

[5] A. Frieze, A. J. Radcliffe, S. Suen, Analysis of a simple greedy matching algorithm on random cubic graphs, in: Proc. 4th Ann. Symposium on Discrete Algorithms (ACM Press, 1993) 341-351.

[6] M. R. Garey and D. S. Johnson, Computers and Intractability - A Guide to the Theory of NP-Completeness (Freeman, New York, 1979).

[7] R. M. Karp, The probabilistic analysis of some combinatorial search algorithms, in: J. F. Traub, ed., Algorithms and Complexity: New Directions and Recent Results (Academic Press, New York, 1976) 1-19.

(32)

[9] T. G. Kurtz, Solutions of ordinary differential equations as limits of pure jump Markov processes, J. Appl. Prob. 7 (1970) 49-58.

[10] C. Lund and M. Yannakakis, On the hardness of approximating minimization prob-lems, in: Proc. STOC’93 (ACM Press, 1993) 286-293.

Références

Documents relatifs

Therefore, the level set based numerical algorithm can perform topology optimization either if the number of holes in the initial structure is sufficiently large [14,17,20] or if

In this paper, we extend set inversion to -algorithms, and apply the resulting method- ology to the guaranteed localization of a robot during a static phase, based on a …nite number

Among the problems that remain to be solved, we need (1) to find symbolic methods for the decomposition of a chain into links with low dimensions; (2) to extend the approach

It was shown to be piecewise linear (in the coecient's space), with a nite number of segments, as a consequence of the niteness of D. Thus when D is not nite, it might occur

ln this section, the impact of each phase of GRASP (greedy randomized con- struction, local search, intensification with path relinking) is examined with re- gard to the best version

For instance, when tackling max independent set in graphs of maximum degree, say, at least 10, a simple tree-search based algorithm with a branching rule of the form either don’t take

Section 2 presents the frictionless contact model and gives four different friction conditions depend- ing on the chosen model: Tresca, Coulomb, Norton Hoff and normal

We first consider the problem of solving total variation denoising for a non decomposable function using active-set methods in Section 5.1, specifically cut functions.. Here,