• Aucun résultat trouvé

Adaptive tests of homogeneity for a Poisson process

N/A
N/A
Protected

Academic year: 2022

Partager "Adaptive tests of homogeneity for a Poisson process"

Copied!
38
0
0

Texte intégral

(1)

www.imstat.org/aihp 2011, Vol. 47, No. 1, 176–213

DOI:10.1214/10-AIHP367

© Association des Publications de l’Institut Henri Poincaré, 2011

Adaptive tests of homogeneity for a Poisson process

M. Fromont

a

, B. Laurent

b

and P. Reynaud-Bouret

c

aUniversité Européenne de Bretagne, CREST-ENSAI, IRMAR, Rennes, France. E-mail:mfromont@ensai.fr bInstitut de Mathématiques de Toulouse, INSA de Toulouse, Université de Toulouse, Toulouse, France.

E-mail:Beatrice.Laurent@insa-toulouse.fr

cCNRS ENS Paris et Université de Nice Sophia-Antipolis, Paris, France. E-mail:reynaudb@unice.fr Received 15 May 2009; revised 8 March 2010; accepted 6 April 2010

Abstract. We propose to test the homogeneity of a Poisson process observed on a finite interval. In this framework, we first provide lower bounds for the uniform separation rates inL2-norm over classical Besov bodies and weak Besov bodies. Surprisingly, the obtained lower bounds over weak Besov bodies coincide with the minimax estimation rates over such classes. Then we construct non-asymptotic and non-parametric testing procedures that are adaptive in the sense that they achieve, up to a possible logarithmic factor, the optimal uniform separation rates over various Besov bodies simultaneously. These procedures are based on model selection and thresholding methods. We finally complete our theoretical study with a Monte Carlo evaluation of the power of our tests under various alternatives.

Résumé. Nous proposons de tester l’homogénéité d’un processus de Poisson observé sur un intervalle borné. Nous établissons tout d’abord des bornes inférieures pour les vitesses de séparation uniformes relativement à la normeL2sur des Besov bodies classiques ou faibles. De façon surprenante, nous obtenons des bornes inférieures sur les Besov bodies faibles qui coïncident avec les vitesses minimax d’estimation sur ce type de classe. Ensuite, nous construisons des procédures de tests non asymptotiques et non paramétriques qui sont adaptatives, au sens où elles atteignent, à un facteur logarithmique près dans certains cas, les vitesses de séparation optimales sur plusieurs classes d’alternatives simultanément. Ces procédures sont basées sur des méthodes de sélection de modèles et de seuillage. Enfin, nous complétons cette étude théorique par des simulations afin d’estimer par la méthode de Monte Carlo la puissance de nos tests sous diverses alternatives.

MSC:Primary 62G10; secondary 62G20

Keywords:Poisson process; adaptive hypotheses testing; uniform separation rate; minimax separation rate; model selection; thresholding rule

1. Introduction

Poisson processes have been used for many years to model a great variety of situations: machine breakdowns, phone calls, etc. Recently Poisson processes become popular for modeling occurrences of words or motifs on the DNA sequence (see [24]). In this context, it is particularly important to be able to detect abnormal behaviors.

With such applications in mind, we consider in this paper the question of testing the homogeneity of a Poisson processN. Since we can only observe a finite number of points of the process, this question has a sense only on a finite interval. For the sake of simplicity, we assume that the Poisson processNis observed on the fixed set[0,1], and that it has an intensityswith respect to some measureμon[0,1]with dμ(x)=Ldx.

Denoting byS0 the set of constant functions on[0,1], our aim is consequently to test the null hypothesis(H0)

“s∈S0,” against the alternative(H1)“s /∈S0.”

This problem of testing the homogeneity of a Poisson process has been widely investigated both from a theoretical and practical point of view (see [2] or [7] for a survey and [5] for a more recent work). In these papers, the alternative

(2)

intensities are monotonous. Another related topic is the problem of testing the simple hypothesis that a point process is a Poisson process with a given intensity. We can cite for instance the papers by Fazli and Kutoyants [12] where the alternative is also a Poisson process with a known intensity, Fazli [11] where the alternatives are Poisson processes with one-sided parametric intensities or Dachian and Kutoyants [10], where the alternatives are self-exciting point processes. The paper by Ingster and Kutoyants [18] is the closest one to the present work. The alternatives considered by Ingster and Kutoyants are Poisson processes with nonparametric intensities in a Sobolev or Besov B(δ)2,q(R)ball with 1≤q <+∞and known smoothness parameterδ.

However, in some practical cases like the study of occurrences of words or motifs on a DNA sequence, such smooth alternatives cannot be considered. The intensity of the Poisson process in these cases may burst at a particular position of special interest for the biologist (see [14] for more details). The question of testing the homogeneity of a Poisson process then becomes “how can we distinguish a Poisson process with constant intensity from a Poisson process whose intensity has some small localized spikes?.” This question has already been partially considered in the seventies in a precursory work by Watson [27]: he proposed a test based on the estimation of the Fourier coefficients of the intensity without evaluating the power of the resulting procedure.

In this paper, we focus on constructing adaptive testing procedures, i.e. which do not use any prior information about the smoothness of the intensitys, but which however have the best possible performances (in a minimax sense).

From a theoretical point of view, we evaluate the performances of the tests in terms of uniform separation rates with respect to some prescribed distanced over various classes of functions. Givenβ∈ ]0,1[, a class of functionsS1, and a levelαtestΦα with values in{0,1}(rejecting(H0)whenΦα=1), the uniform separation rateρ(Φα,S1, β) ofΦα over the classS1is defined as the smallest positive numberρsuch that the test has an error of second kind at most equal toβfor all alternativessinS1at anL2-distanceρfromS0. More precisely, ifPs denotes the distribution of the Poisson processNwith intensitys,

ρ(Φα,S1, β)=inf

ρ >0, sup

s∈S1,d(s,S0)>ρ

Psα=0)≤β

(1.1)

=inf

ρ >0, inf

s∈S1,d(s,S0)>ρPsα=1)≥1−β

. (1.2)

In view of the practical situations of our interest, we study some classes of alternatives that can be very irregular, for instance that can have some localized spikes. We then consider some classical Besov bodies and also some spaces that can be viewed as weak versions of these classical Besov bodies and that are defined precisely in the following.

The interested reader may find in [23] some illustrations of functions in weak Besov spaces and how the smoothness parameters of the functions govern the proportion and amplitude of their spikes.

As a first step, we evaluate the best possible value of the uniform separation rate over these spaces. In other words, we give a lower bound for

ρ(S1, α, β)=inf

Φα

ρ(Φα,S1, β), (1.3)

where the infimum is taken over all level αtestsΦα, and whereS1 can be either a Besov body or a weak Besov body. This quantity introduced by Baraud [3] as the(α, β)-minimax rate of testing over S1 or the minimax sepa- ration rate over S1 is a stronger version of the (asymptotic) minimax rate of testing usually considered. The key reference for the computation of minimax rates of testing in various statistical models is the series of papers due to Ingster [16]. Concerning the Poisson model, Ingster and Kutoyants [18] give the minimax rate of testing for Sobolev or BesovB(δ)2,q(R)balls with 1≤q <+∞and smoothness parameter δ >0. They find that this rate of testing for the Sobolev or Besov norm or semi-norm is of orderL2δ/(4δ+1). Let us note that we find here lower bounds for the classical Besov bodies similar to Ingster and Kutoyants’ ones. Furthermore, our lower bounds for the weak Besov bodies are larger than the ones for classical Besov bodies. Alternatives in weak Besov bodies are in fact so irreg- ular that it is as difficult to detect them as to estimate them. The problem of estimation in weak Besov spaces is solved by using thresholding procedures: indeed the weak Besov spaces are closely related to the maxisets of those procedures (see [19] in the Gaussian framework and [22] in a Poisson model). To our knowledge, no previous re- sults of this kind exist for weak Besov bodies in testing problems, even in more classical statistical models, like the density model. Despite the similarity of both models, our lower bounds over weak Besov bodies cannot however

(3)

be straightly transposed to the density model since our proofs heavily rely on the Poissonian independence proper- ties.

As a second and main step, we construct non asymptotic levelαtests which achieve, up to a possible logarithmic factor, the minimax separation rates over many Besov bodies and weak Besov bodies simultaneously, whereas using no prior information about the smoothness of the intensity s. Our idea here is to combine some model selection methods that are effective for alternatives in classical Besov bodies and a thresholding type approach, inspired by the thresholding rules used for adaptive estimation in weak Besov bodies. Key tools in the proofs of our results are exponential inequalities for U-statistics of order 2 due to Houdré and Reynaud-Bouret [15].

Of course, both model selection and thresholding approaches have already been used to construct adaptive tests in various statistical models. One can cite among others the papers by Spokoiny [25,26] in Gaussian white noise models or by Baraud, Huet and Laurent [4] in a Gaussian regression framework. These papers propose adaptive tests which combine methods closely related to both model selection and thresholding ones. As for the density framework, adaptive tests were proposed by Ingster [17] or Fromont and Laurent [13], using model selection type methods and by Butucea and Tribouley [6] using thresholding type methods.

The present work is organized as follows. In Section2, we provide lower bounds for the minimax separation rates over various Besov bodies. Our testing procedures are defined in Section3, and their uniform separation rates over Besov bodies are established in Section4. We carry out a simulation study in Section5to illustrate these theoretical results, and the proofs are postponed to the last section.

2. Lower bounds for the minimax separation rates over Besov bodies

We consider the Poisson processNwith intensityswith respect to some measureμon[0,1], with dμ(x)=Ldx. In the following, we assume thatsbelongs toL2([0,1]), and·,·, · andd respectively denote the scalar product

f, g =

[0,1]f (x)g(x)dx, theL2-norm

f2=

[0,1]f2(x)dx, and the associated distance.

Let us denote the Haar basis ofL2([0,1])by{φ0, φ(j,k), j ∈N, k∈ {0, . . . ,2j−1}}with φ0(x)=1[0,1](x)

and

φ(j,k)(x)=2j/2ψ

2jxk

, (2.1)

whereψ (x)=1[0,1/2[(x)1[1/2,1[(x).

We setα0= s, φ0and for everyj∈N, k∈ {0, . . . ,2j−1},α(j,k)= s, φ(j,k). We can now introduce the Besov bodies defined forδ >0,R >0 by

Bδ2,(R)=

s≥0, s∈L2 [0,1]

, s=α0φ0+

j∈N 2j1

k=0

α(j,k)φ(j,k),

j∈N,

2j1 k=0

α2(j,k)R222j δ , (2.2)

(4)

and more generally forp≥1,R >0 andδ >max(0,1/p−1/2), by Bp,δ(R)=

s≥0, s∈L2 [0,1]

, s=α0φ0+

j∈N 2j1

k=0

α(j,k)φ(j,k),

j∈N,

2j1 k=0

|α(j,k)|pRp2pj (δ+1/21/p) . (2.3)

As in [22], we also introduce some weaker versions of the above Besov bodies given forγ >0 andR>0 by

Wγ R

=

s≥0, s∈L2 [0,1]

, s=α0φ0+

j∈N 2j1

k=0

α(j,k)φ(j,k),

t >0,

j∈N 2j1

k=0

α(j,k)2 1α2

(j,k)tR2t2γ /(1+2γ ) . (2.4)

Fixing some levels of error αandβ in ]0,1[, and denoting byL(R)the set of functions bounded by R, our purpose in this section is to find sharp lower bounds forρ(B2,δ(R)Wγ(R)∩L(R), α, β), whereρ is defined by (1.3).

Starting from a general idea developed by Ingster [16], we obtain the following result.

Theorem 1. Assume thatR >0,R>0,andR≥2,and fix some levelsαandβ in]0,1[such thatα+β≤0.59.

Let us consider the different cases that are represented in Fig.1.

In case(i)whereδ≥max(γ /2, γ /(1+2γ )),then lim inf

L→+∞L2δ/(1+4δ)ρ

B2,δ(R)Wγ R

∩L R

, α, β

>0.

In case(ii)whereδ < γ /2andγ >1/2,then lim inf

L→+∞

L lnL

γ /(1+2γ )

ρ

B2,δ(R)Wγ R

∩L R

, α, β

>0.

Fig. 1. The set of possible parameters (δ, γ): visualization of the different cases appearing in Theorem1.

(5)

In case(iii)whereδ < γ /(1+2γ )andγ≤1/2,then lim inf

L→+∞

L1/4L2γ /((1+4δ)(1+2γ )) ρ

Bδ2,(R)Wγ R

∩L R

, α, β

>0.

Comments:

(1) For the whole set of parameters(δ, γ )such thatδγ /(1+2γ ), we prove in Section4that these lower bounds are actually sharp.

(2) We have in case (i) lower bounds which coincide with the minimax rates of testing obtained by Ingster and Kutoyants [18] when testing that a periodic Poisson process has a given intensity in the Besov spacesB2,(δ)(R).

We know (see [13,17], for instance) that such rates can be achieved by some multiple testing procedure based on model selection type methods. This is the principle of our first procedure described in Section3.1.

(3) We notice that the lower bounds obtained in case (ii) are equal to the minimax estimation rates on the maxisets of the thresholding estimation procedure, namelyBκγ /(12, +2γ )(R)Wγ(R)with some constantκ <1 (see [19, 22,23] for more details). This means that it is at least as difficult to test as to estimate over such classes of functions, phenomenon which is quite unusual. Since the minimax estimation rates on these classes are achieved by thresholding rules, it will be natural to construct a testing procedure based on thresholding methods: this is the idea that originated our second procedure described in Section3.2.

(4) As for the case (iii), our present results do not allow us to say anything about the possible optimality of the obtained lower bound.

3. Two tests of homogeneity

Let us recall thatS0denotes the set of constant functions on[0,1]and that we assume thatsbelongs toL2([0,1]).

In this section, we construct levelαtests of the null hypothesis(H0)“s∈S0,” against the alternative(H1)“s /∈S0,”

from the observation of the Poisson processN, or the points{Xl, l=1, . . . , NL}of the Poisson process.

We introduce two testing procedures that come from two different statistical approaches. The first one originates in general model selection methods, while the second one is closer to the thresholding type methods.

In order to understand the global ideas of these procedures, let us notice that the squaredL2-distanced2(s,S0) betweensand the set of constant functionsS0can be rewritten as

d2(s,S0)=

[0,1]

s(x)

[0,1]s(y)dy 2

dx

= s2α20

=

λΛ

α2λ,

whereα0= s, φ0, and for allλΛ= {(j, k), j∈N, k∈ {0, . . . ,2j−1}},αλ= s, φλ. For everyλΛ,αλcan be estimated by

αλ= 1

L

[0,1]φλ(x)dNx, which is also equal to

αλ= 1

L

NL

l=1

φλ(Xl).

From this variable, we deduce an unbiased estimator ofα2λgiven by Tλ=αλ2− 1

L2

[0,1]φλ2(x)dNx= 1 L2

NL

l=l=1

φλ(Xlλ(Xl). (3.1)

(6)

Our first approach will consist in constructing estimators ofd2(s,S0)=

λΛα2λ based on a combination of the Tλ’s, and in rejecting the null hypothesis when one of these estimators is too large. This was already the spirit of Watson’s procedure (see [27]). Our second approach is related to the test considered in [4] to detect local alternatives.

It will consist in considering a set ofTλ’s and rejecting the null hypothesis directly when one of theTλ’s is too large.

Let us now precisely define both procedures.

3.1. A first procedure based on model selection

Assuming that s∈L2([0,1]), a natural idea is to construct a testing procedure from an estimation of the squared L2-distanced2(s,S0).

In order to estimate this functional of s, following the ideas of Laurent [20] and Fromont and Laurent [13], we introduce embedded finite-dimensional linear subspaces of L2([0,1]). We choose here to consider forJ ≥1 the subspacesSJ generated by the subsets{φ0, φλ, λΛJ}of the Haar basis defined by (2.1), withΛJ = {(j, k),j∈ {0, . . . , J−1},k∈ {0, . . . ,2j−1}}. Each subspaceSJ is called amodel. We denote byDJ=2Jthe dimension ofSJ, and bysJ the orthogonal projection ofsonto the modelSJ.

Focusing on one model SJ, we estimate d2(s,S0)= s2α02 by the unbiased estimator of sJ2α02=

λΛJαλ2given by TJ =

λΛJ

Tλ, (3.2)

withTλdefined by (3.1). The estimatorTJ obviously depends on the choice of the modelSJ.

Since we do not want to choose a priori such a model, we consider a collection of models{SJ, JJ}whereJ is a finite subset ofN, and the corresponding collection of estimators{TJ, JJ}.

The procedure that we introduce here then consists in rejecting(H0)“s∈S0” when there existsJ inJ such that the estimatorTJ given by (3.2) is too large.

At this point there are several ways to decide whenTJ is too large.

In all cases, we use the well-known argument that, conditionally on the event “the number of pointsNL falling into [0,1] isn,” the points of the process obey the same law as ann-sample (X˜1, . . . ,X˜n)with common density s/

[0,1]s(x)dx. It follows that for alln∈N, Ps

TJ > q|NL=n

=P

1 L2

λΛJ

n l=l=1

φλ(X˜lλ(X˜l) > q

.

Under the null hypothesis, the intensitysis constant on[0,1], and theX˜l’s are i.i.d., with uniform distribution on [0,1]. This distribution is free from the parameters. As a consequence, for everyu∈ ]0,1[, we can introduce and estimate by Monte Carlo experiments the(1u)quantile of the distribution ofTJ|NL=nunder the null hypothesis, that we denote byqJ(n)(u).

We now consider the test statistics:

Tα(1)=sup

J∈J

TJqJ(NL) uJ,α(NL)

, (3.3)

withuJ,α(NL)to be correctly chosen.

Finally, we define the corresponding test function:

Φα(1)=1T(1)

α >0. (3.4)

And our first test consists in rejecting the null hypothesis(H0)whenΦα(1)=1.

Let us see how we can chooseuJ,α(NL)so that our test has a levelα.

(7)

An obvious possibility is to set uJ,α(n)= α

|J| for everyJinJ andninN.

This choice corresponds to a Bonferroni procedure andΦα(1)actually defines a levelαtest. Indeed, forsS0, Ps

sup

J∈J

TJqJ(NL) uJ,α(NL)

>0

=

n∈N

Ps

sup

J∈J

TJqJ(n) α

|J|

>0NL=n

Ps(NL=n)

n∈N

J∈J

α

|J|Ps(NL=n)

α.

Another choice foruJ,α(n), inspired by Fromont and Laurent [13], consists in setting uJ,α(n)=eWJsup

u∈ ]0,1[,sup

s∈S0

Ps

sup

J∈J

TJqJ(n)

ueWJ

>0NL=nα

, (3.5)

where{WJ, JJ}is a collection of positive weights such that

J∈J

eWJ ≤1.

Since this new definition ofuJ,α(n)leads to a less conservative procedure, this is the one that we use in the rest of the paper. We notice thatuJ,α(n)αeWJ for everyninN, and that in practice,uJ,α(n)can be estimated by Monte Carlo experiments.

3.2. A second procedure based on a thresholding approach

Let us recall here that the squaredL2-distanced2(s,S0)betweensand the set of constant functionsS0is equal to

λΛαλ2 and that Tλ defined by (3.1) is an unbiased estimator of α2λ. Based on general thresholding ideas, our second procedure consists in fixing someJ¯≥1 and rejecting the null hypothesis(H0)when there existsλinΛJ¯such thatTλis too large.

Let us now see what we mean by “Tλis too large.” We can still use the fact that Ps(Tλ> q|NL=n)=P

1 L2

n l=l=1

φλ(X˜lλ(X˜l) > q

,

and that under the null hypothesis, theX˜l’s are i.i.d., with uniform distribution on[0,1]. We can therefore introduce and estimate by Monte Carlo experiments the(1u)quantile of the distribution ofTλ|NL=nunder the null hy- pothesis, that we denote byqλ(n)(u), for everyu∈ ]0,1[. Notice that forλ=(j, k)ΛJ¯,qλ(n)(u)does not depend onk.

We set Tα(2)= sup

λΛJ¯

Tλqλ(NL) u(Nλ,αL)

, (3.6)

withu(Nλ,αL)to be correctly chosen.

(8)

We also define Φα(2)=1T(2)

α >0. (3.7)

Our test consists in rejecting the null hypothesis(H0)whenΦα(2)=1.

Let us now see how we chooseu(n)λ,α. An obvious choice corresponding to the Bonferroni procedure would be u(n)λ,α= α

2jJ¯ for everyλ=(j, k)inΛJ¯andninN. To obtain a less conservative procedure, we prefer setting

u(n)λ,α=u(n)α

2jJ¯ for everyλ=(j, k)inΛJ¯andninN, (3.8)

with

u(n)α =sup

u∈ ]0,1[,sup

s∈S0

Ps

sup

(j,k)ΛJ¯

T(j,k)q(j,k)(n) u

2jJ¯

>0NL=n

α

. (3.9)

WhensS0, Ps

sup

λΛJ¯

Tλqλ(NL) u(Nλ,αL)

>0

=Ps

sup

(j,k)ΛJ¯

T(j,k)q(j,k)(NL) u(Nα L)

2jJ¯

>0

=

n∈N

Ps

sup

(j,k)ΛJ¯

T(j,k)q(j,k)(n) u(n)α

2jJ¯

>0NL=n

P(NL=n)

α,

which means thatΦα(2)defines a levelαtest.

Note thatu(n)α satisfiesu(n)ααand that it can be estimated in practice by Monte Carlo experiments.

Comments:

(1) Though the two testing procedures defined by (3.4) and (3.7) are very different in their spirit, they can formally be written in a common way. For any subsetΛofΛ, we denote bySΛthe subspace generated by{φ0, φλ, λΛ}, bysΛ the orthogonal projection ofsontoSΛ, and we introduce the unbiased estimatorTΛ=

λΛTλ of sΛ2α02=

λΛαλ2. Then our test functions can be written as

Φα=1Tα>0, (3.10)

where Tα=sup

Λ∈C

TΛtΛ,α (NL) ,

andCis a finite collection of subsets ofΛ. Noticing thatTJ=TΛ

J, we can easily see that our first test amounts in taking a collectionCequal to{ΛJ, JJ}, andtΛ

J

(NL)=qJ(NL)(uJ,α(NL)). Furthermore, our second test amounts in taking a collectionCcomposed of all subsets ofΛJ¯, and forΛΛJ¯,tΛ,α (NL)=

λΛqλ(NL)(u(Nλ,αL)). Indeed, there exists a subsetΛofΛJ¯such that

λΛTλ>

λΛqλ(NL)(u(Nλ,αL))if and only if there existsλinΛJ¯such thatTλ> qλ(NL)(u(Nλ,αL)).

(9)

Such a common expression will be particularly useful to derive the properties of the tests.

It also allows us to see our tests as multiple testing procedures. Indeed, we can consider that for eachΛinC, we construct a test rejecting the null hypothesis whenTΛtΛ,α(NL)>0. We thus obtain a collection of tests and we finally decide to reject the null hypothesis when it is rejected for at least one of the tests of the collection.

(2) Both procedures have a specific interest to prove the optimality of lower bounds obtained in Theorem1. We will actually prove in the next section that the first one achieves the lower bounds obtained in case (i) of Theorem1(up to a possible logarithmic factor) whereas the second one achieves lower bounds obtained in case (ii) of Theorem1.

However, if we want a procedure that achieves lower bounds of cases (i) and (ii) simultaneously, we will have to consider the test which consists in mixing the two procedures. In this case, we reject the null hypothesis(H0) when sup{Φα/2(1), Φα/2(2)} =1.

4. Uniform separation rates

In this section, we evaluate the performances of our new testing procedures from a theoretical point of view. More precisely, we prove that our procedures are optimal in the sense that their uniform separation rates over Besov bodies are of the same order as the lower bounds forρ obtained in Section2. These results justify the construction of our procedures as well as they provide the upper bounds needed for the exact evaluation of the minimax separation rates over weak and classical Besov bodies in the Poisson framework.

In the following, the expressionC(α, β, R, R, R, δ, γ , . . .)orCk(α, β, R, R, R, δ, γ , . . .)is used to denote some constant which only depends on the parametersα, β, R, R, R, δ, γ , . . . ,and which may vary from line to line.

4.1. Uniform separation rates of the first procedure 4.1.1. The error of second kind

The aim of the following theorem is to give a condition on the alternative so that our first levelαtest has a prescribed error of second kind.

Theorem 2. Assume thats∈L([0,1]),and thatL≥1.Fix some levelsαandβ in]0,1[,and letΦα(1) be the test function defined by(3.4).There exist some positive constantsC1(α, β,||s||),C2(α, β),C3,C4(β)andC5such that whenssatisfies

d2(s,S0) > inf

J∈J

||ssJ||2+C1

α, β,||s||DJ

L +C2(α, β)DJ L2 +

C3

[0,1]s(x)dx+C4(β)

DJWJ L +WJ

L

+C5DJWJ2 L2

(4.1) then

Ps

Φα(1)=0

β.

Comment:Considering here a multiple testing procedure instead of a simple one allows to obtain in the right-hand side of the inequality(4.1)an infimum over allJ inJ at the only price of introducing some terms withWJ. These last terms will appear in the following uniform separation rates over classical Besov bodies as a ln lnLfactor, which is now known to be the price to pay for adaptivity in some classical statistical models. As a consequence, our multiple testing procedure will be proved to be adaptive over classical Besov bodies in Proposition1, which would not occur with a simple testing procedure.

4.1.2. Uniform separation rates over Besov bodies

In this section, we evaluate the uniform separation ratesρ(Φα(1),Bδ2,(R)∩L(R), β)whereρis defined by (1.1), andB2,δ(R)is any Besov body defined by(2.2).

(10)

Let us first notice that the functions ofBδ2,(R)are well approximated by their projections onto subspaces of the collection{SJ, JJ}considered in our first procedure, in the sense that ifsB2,δ(R), then

ssJ2c(δ)R2DJ.

As a consequence we can use Theorem2to obtain upper bounds for the uniform separation rates of our test.

We denote byxthe integer part ofx.

Proposition 1. Assume thatln lnL≥1. Given some levelsαand β in]0,1[,letΦα(1) be the test function defined by(3.4)withJ = {1, . . . ,log2(L2/(ln lnL)3)}andWJ =ln|J|for everyJ inJ.

For everyδ >0,R >0andR>0,there exists some positive constantC(α, β, R, δ)such that whensbelongs to B2,δ(R)∩L(R)and satisfies

d2(s,S0) > C

α, β, R, δ

R2/(4δ+1)

ln lnL L

4δ/(4δ+1)

+R2

(ln lnL)3 L2

+ln lnL L

, then

Ps

Φα(1)=0

β.

In particular,there exist some positive constantsL0(δ)andC(α, β, R, R, δ)such that ifL > L0(δ),then ρ

Φα(1),Bδ2,(R)∩L R

, β

C

α, β, R, R, δ√ ln lnL

L

2δ/(4δ+1)

. Comments:

(1) Our first testing procedure is therefore adaptive: indeed, for largeL, it achieves the lower bounds for the minimax separation rates over all the spacesB2,δ(R)Wγ(R)∩L(R)withδ≥max(γ /2, γ /(1+2γ ))simultaneously up to a possible ln lnLfactor (see Theorem1).

However it does not achieve the optimal separation rates obtained in the case whereδ < γ /2 andγ >1/2. In this range of parameters, the regularity inγis higher than the regularity inδ, meaning that the weak Besov body governs the separation rate. That is the reason why we introduced the thresholding type procedure.

(2) The upper bound for the uniform separation rate obtained here is exactly of the same order as the (asymptotic) adaptive minimax rate of testing obtained by Ingster [17] in the density model, replacing the numbernof obser- vations in the density model by the parameterLof the Poisson model. In particular the ln lnn, replaced here by a ln lnLfactor, is proved to be necessary in the density model for adaptive procedures.

(3) It is easy to see that Bδp,(R)Bδ2,(R)when p >2. So the result of Proposition 1 directly leads to upper bounds for the uniform separation ratesρ(Φα(1),Bδp,(R)∩L(R), β)whenp >2. These rates, obtained in the Poisson framework, correspond to the ones in some Gaussian models (see [25], for instance) or in the density model (see [17]).

(4) Note that one could also consider some tests based on the Fourier basis as well as the Haar basis, as Fromont and Laurent [13] did in the density model. The theoretical results would remain unchanged, and the practical performances of the procedure would be better when considering smooth alternatives (see [13] for more details and Section5). We have only considered here tests based on the Haar basis for the sake of simplicity.

4.2. Uniform separation rates of the second procedure 4.2.1. The error of second kind

From the common expression (3.10) of the test function for the two procedures, we obtain here a result similar to Theorem2for the error of second kind of our second test.

(11)

Theorem 3. Assume thats∈L([0,1]),and thatL≥1.Fix some levelsαandβ in]0,1[,and letΦα(2) be the test function defined by(3.7).Recall that for any subsetΛofΛ,SΛandsΛrespectively denote the subspace generated by{φ0, φλ, λΛ}and the orthogonal projection ofsontoSΛ.Denoting byDΛthe dimension ofSΛ,there exist some positive constantsC1(α, β,||s||),C2(β),C3(α),C4(α, β)andC5(α)such that whenssatisfies

d2(s,S0) > inf

ΛΛJ¯

||ssΛ||2+C1

α, β,||s||DΛ L +2J /2¯

L3/2

+C2(β)2J¯ L2 +

C3(α)

[0,1]s(x)dx+C4(α, β)

DΛln(2J¯J )¯

L +C5(α)DΛ2J¯ln2(2J¯J )¯ L2

(4.2) then

Ps

Φα(2)=0

β.

4.2.2. Uniform separation rates over Besov bodies

Proposition 2. Assume thatlnL≥1.Given some levelsαandβ in]0,1[,letΦα(2) be the test defined by(3.7)with J¯= log2(L/lnL).

For everyδ >0andγ >0,R >0,R>0andR>0,there exists some positive constantC(α, β, R, δ, γ )such that ifsbelongs toB2,δ(R)Wγ(R)∩L(R)and satisfies

d2(s,S0) > C

α, β, R, δ, γlnL L +R2

lnL L

+R2 lnL

L

2γ /(1+2γ )

+R2+ lnL

L

, then

Ps

Φα(2)=0

β.

In particular,whenδγ /(1+2γ ),there exist some positive constantsL0(δ, γ )andC(α, β, R, R, R, δ, γ )such that ifL > L0(δ, γ ),then

ρ

Φα(2),Bδ2,(R)Wγ R

∩L R

, β

C

α, β, R, R, R, δ, γlnL L

γ /(1+2γ )

. Comments:

(1) Our second testing procedure is still adaptive in the sense that for large L, it achieves the lower bounds for the minimax separation rates over all the spaces B2,δ(R)Wγ(R)∩L(R) withγ /(1+2γ )≤δ < γ /2 simultaneously (see Theorem1). In this case, we also remark that these rates are so large that there is no further price to pay for adaptivity in the sense that the upper bound does not involve any extra logarithmic factor. To our knowledge, this phenomenon is completely new for nonparametric testing procedures.

(2) Our second procedure achieves the lower bounds for the minimax separation rates over all the spacesB2,δ(R)Wγ(R)∩L(R)withγ /(1+2γ )≤δ < γ /2 simultaneously, but it does not whenδ≥max(γ /2, γ /(1+2γ )).

To obtain a test that achieves the minimax separation rates in both cases, our two procedures need to be combined.

4.3. Uniform separation rates of the combined procedure

Corollary 1. Assume thatln lnL≥1.Fix some levelsαandβin]0,1[.LetΦα/2(1) be the levelα/2test defined by(3.4) withJ = {1, . . . ,log2(L2/(ln lnL)3)}andWJ =ln|J| for everyJ inJ.LetΦα/2(2) be the levelα/2test defined by(3.7)withJ¯= log2(L/lnL).We considerΦα(3)=sup{Φα/2(1), Φα/2(2)}.

(12)

(i) For all δ >0 and γ >0, R >0, R >0 and R >0, there exist some positive constants L0(δ) and C(α, β, R, R, δ)such that ifL > L0(δ),then

ρ

Φα(3),Bδ2,(R)Wγ R

∩L R

, β

C

α, β, R, R, δ√ ln lnL

L

2δ/(4δ+1)

.

(ii) For all(δ, γ )such thatδγ /(2γ+1),R >0,R>0andR>0,there exist some positive constantsL0(δ, γ ) andC(α, β, R, R, R, δ, γ )such that ifL > L0(δ, γ ),then

ρ

Φα(3),Bδ2,(R)Wγ

R

∩L R

, β

C

α, β, R, R, R, δ, γlnL L

γ /(1+2γ )

. Comments:Since

Ps

Φα(3)=0

≤inf Ps

Φα/2(1) =0 ,Ps

Φα/2(2) =0 , the proof of this result directly comes from Propositions1and2.

The upper bounds for the uniform separation rates of the above combined procedure over B2,δ(R)Wγ(R)∩ L(R)obtained here actually coincide with the lower bounds of Theorem1for the whole set of parameters(δ, γ ) such thatδγ /(2γ+1)(up to a ln lnLfactor whenδγ /2). This proves that our combined procedure is conse- quently adaptive over the spaces Bδ2,(R)Wγ(R)∩L(R), and also that the lower bounds of Theorem1are sharp for this set of parameters.

We however do not obtain such a result for the set of parameters(δ, γ )such thatδ < γ /(2γ+1). Hence we do not know the exact order of the uniform separation rate of our combined procedure overB2,δ(R)Wγ(R)∩L(R) whenδ < γ /(2γ+1), and we cannot say here whether our lower bounds for these spaces are optimal or not.

5. Simulation study

We aim in this section at studying the performances of our tests from a practical point of view. We consider several intensitiessdefined on[0,1]such that1

0s(x)dx=1.N denotes here a Poisson process with intensityLs on[0,1] with respect to the Lebesgue measure, andPs the distribution of this process. We denote bys0the intensity which is constant (equal to 1) on[0,1]. We chooseL=100 and a level of testα=0.05.

Let us now recall that our first procedure may be based on the test statistics Tα(1)=sup

J∈J

TJqJ(NL)

uα(NL)/|J| ,

whereqJ(n)(u)denotes the(1u)quantile ofTJ|NL=nunder the hypothesis thats=s0, anduα(n)is chosen such that

uα(n)=sup

u∈ ]0,1[,Ps0

sup

J∈J

TJqJ(n) u/|J|

>0NL=nα

. The null hypothesis(H0)“s=s0” is rejected whenTα(1)>0.

We chooseJ = {1, . . . ,6}. For 40≤n≤160, we estimate the quantitiesuα(n)and the quantilesqJ(n)(uα(n)/|J|) for allJinJ. These estimations are based on the simulation of 200 000 independent samples with sizen, uniformly distributed on [0, 1]. Half of the samples is used to estimate the quantilesqJ(n)(u/|J|)foruvarying on a grid over [0,1], and the other samples are used to estimate the probabilities occurring in the definition ofuα(n). Finally,uα(n)is estimated by the largest value on the grid such that these estimated probabilities are smaller thanα.

Let us also recall that our second procedure is based on the test statistics Tα(2)= sup

λΛJ¯

Tλqλ(NL) u(Nλ,αL)

,

Références

Documents relatifs

While several approaches make it possible to relate high level extra-functional properties to low-level concerns [11, 12, 18, 9], none of them take into account adaptable

Suite à l'indisponibilité de la bibliothèque rlu capterur de glucose dans Proteus, on a décidé de remplacer ce capteur par des sorurce de courant. On a aussi en

The approach consists in appropriately writing the problem in the wavelet domain and then selecting both the block size and threshold parameter at each resolution level by

As said previously, Juditsky and Lambert-Lacroix [24] established minimax rates for the problem of estimating a density with an infinite support for the particular class of Besov

Namely, Ingster, Tsybakov and Verzelen (2010), Arias-Castro, Cand` es and Plan (2011) deal with an asymptotic setting under additional assumptions on the parameters of the problem

Still adopting the minimax point of view, we will see that when considering the transitory change detection problem, adaptation to both change location and length has a

The test proposed by Eubank and Hart (1992) is based on a penalized criterion which is related to the well-known Mallows C p and has the property to achieve the parametric rate

In Section 2, we will state some general conditions under which the navigation tree converges to the directed navigation forest for the local weak convergence on graphs as it defined