• Aucun résultat trouvé

Proofs of the probabilistic results for Hawkes processes

Dans le document 1.3. Notation and overview of the paper (Page 43-52)

6. Simulations for the multivariate Hawkes process

7.4. Proofs of the probabilistic results for Hawkes processes

7.4.1. Proof of Lemma1

LetK(n) denote the vector of the number of descendants in then’th generation from a single ancestral point of type`, defineK(0) =e`and letW(n) =Pn

k=0K(k) denote the total number of points in the firstngenerations. Define forθ∈RM

φ`(θ) = logE`eθTK(1).

Thus, φ`(θ) is the log-Laplace transform of the distribution of K(1) given that there is a single initial ancestral point of type `. We define the vector φ(θ) by φ(θ)0 = (φ1(θ), ..., φM(θ)). Note that φ only depends on the law of the number of children per parent, i.e. it only depends on Γ. Then

E`eθTW(n) = E` Definingg(θ) =θ+φ(θ) we arrive by recursion at the formula

E`eθTW(n) = E`eg◦(n−1)(θ)TK(1)+θTW(0)

= eφ(g◦(n−1)(θ))``

= eg◦n(θ)`.

where for anyn, g◦(n) =g◦ · · · ◦g n times. Or, in other words, we have the following representation

logE`eθTW(n)=g◦n(θ)` of the log-Laplace transform ofW(n).

Below we show that φ is a contraction in a neighborhood containing 0, that is, for somer >0 and a constantC <1 (and a suitable norm),||φ(s)|| ≤C||s|| for||s|| ≤r. If θis chosen such that

||θ||

1−C ≤r

we have||θ|| ≤r, and if we assume thatg◦k(θ)∈B(0, r) fork= 1, . . . , n−1 then

||g◦n(θ)|| ≤ ||θ||+||φ(g◦(n−1)(θ))||

≤ ||θ||+C||g◦(n−1)(θ)||

≤ ||θ|| 1 +C+C2+. . .+Cn

≤ r

Thus, by induction, g◦n(θ)∈B(0, r) for all n≥1. Since n7→Wm(n) is increasing and goes toWm(∞) forn→ ∞, withWm(∞) the total number of points in a cluster of type m, and sinceW =P

mWm(∞) =1TW(∞), we have by monotone convergence that for ϑ∈R

logE`eϑW = lim

n→∞g◦n(ϑ1)`.

By the previous result, the right hand side is bounded if |ϑ| is sufficiently small. This completes the proof up to proving thatφis a contraction.

To this end we note thatφis continuously differentiable (onRM in fact, but a neigh-borhood around 0 suffices) with derivativeDφ(0) = Γ at 0. Since the spectral radius of Γ is strictly less than 1 there is aC <1 and, by the Householder theorem, a norm|| · ||

onRM such that for the induced operator norm of Γ we have

||Γ||= max

x:||x||≤1||Γx||< C

Since the norm is continuous andDφ(s) is likewise there is anr >0 such that

||Dφ(s)|| ≤C <1

for||s|| ≤r. This, in turn, implies thatφis Lipschitz continuous in the ballB(0, r) with Lipschitz constantC, and sinceφ(0) = 0 we get

||φ(s)|| ≤C||s||

for||s|| ≤r. This ends the proof of the lemma.

Note that we have not at all used the explicit formula forφabove, which is obtainable and simple since the offspring distributions are Poisson. The only thing we needed was the fact thatφis defined in a neighborhood around 0, thus that the offspring distributions are sufficiently light-tailed.

7.4.2. Proof of Proposition2

We use the cluster representation, and we note that any cluster with ancestral point in [−n−1,−n] must have at least n+ 1− dAepoints in the cluster if any of the points are to fall in [−A,0). This follows from the assumption that all theh(m)` -functions have support in [0,1]. With ˜NA,`the number of points in [−A,0) from a cluster with ancestral points of type` we thus have the bound

A,`≤X

n An

X

k=1

max{Wn,k−n+dAe,0}

where An is the number of ancestral points in [−n−1,−n] of type ` and Wn,k is the number of points in the respective clusters. Here the An’s and the Wn,k’s are all inde-pendent, the An’s are Poisson distributed with mean ν` and the Wn,k’s are i.i.d. with the same distribution asW in Lemma1. Moreover,

Hn`) :=E`eϑ`max{W−n+dAe,0}≤P`(W ≤n− dAe) +e−ϑ`(n−dAe)E`eϑ`W, which is finite for|ϑ`|sufficiently small according to Lemma1. Then we can compute an upper bound on the Laplace transform of ˜NA,`:

Eeϑ`N˜A,` ≤ Y

n

E

An

Y

k=1

E

eϑ`max{Wn,k−n+dAe,0}|An

≤ Y

n

EHn`)An

= Y

n

eν`(Hn`)−1)

= eν`Pn(Hn`)−1) SinceHn`)−1≤e−ϑ`(n−dAe)E`eϑ`W we haveP

n(Hn`)−1)<∞, which shows that the upper bound is finite. To complete the proof, observe thatN[−A,0)=P

`A,`where N˜A,`for`= 1, . . . , M are independent. Since all variables are positive, it is sufficient to takeθ= min`ϑ`.

7.4.3. Proof of Proposition3

In this paragraph, the notation simply denotes a generic positive absolute constant that may change from line to line. The notation θ12,... denotes a positive constant depending onθ1, θ2, . . .that may change from line to line.

Let

u=C1σlog3/2(T)√

T +C2b(log(T))2+η, (7.13)

where the choices ofC1 and C2 will be given later. For any positive integerksuch that

Similarly to [51], we introduce ( ˜Mqx)q a sequence of independent Hawkes processes, each being stationary with intensities per mark given byψt(m). For eachq, we then introduce Mqx the truncated process associated with ˜Mqx, where truncation means that we only consider the points lying in [2qx−A,2qx+x].So, if we set

whereTe represents the time to extinction of the process. More preciselyTe is the last point of the process if in the cluster representation only ancestral points before 0 are appearing. For more details, see section 3 of [51]. So, denoting al the ancestral points with markslandHal

lthe length of the corresponding cluster whose origin isal, we have:

Te= max

where ˜N(l) denotes the process associated with the ancestral points with marksl. So, is the number of points in the cluster. But if all the interaction functions have support in [0,1], one always have thatH0l < W. Hence

Now, let us focus on the first termB of (7.14), where B =P

Let us consider some ˜N where ˜N will be fixed later and let us define the measurable events

We haveP(Ωc)≤P

qP(Ωcq).Each Ωq can also be easily controlled. Indeed it is sufficient to split [2qx−A,2qx+x] in intervals of sizeA(there are aboutα,A,f0log(T) of those) and require that the number of points in each subinterval is smaller than ˜N/2. By stationarity, we obtain that

P(Ωcq)≤α,A,f0log(T)P(N[−A,0)>N˜/2).

Using Proposition2 withu=dN˜/2e+ 1/2, we obtain:

P(Ωcq)≤α,A,f0log(T) exp(−α,A,f0N˜) andP(Ωc)≤α,A,f0Texp(−α,A,f0N˜).

(7.15) Note that this control holds for any positive choice of ˜N. Hence this gives also the following Lemma that will be used later.

Lemma 3. For any R>0,

P there existst∈[0, T]| Mqx|[t−A,t)>R)≤α,A,f0Texp(−α,A,f0R).

Hence by taking ˜N =C3log(T) forC3 large enough this is smaller thanα,A,f0T−α0, whereα0= max(α,2).

It remains to obtain the rate ofD:=P(P

qFq ≥u/2 and Ω). For any positive constant θthat will be chosen later, we have:

D ≤ eθu2 E eθPqFqY

q

1q

!

≤ eθu2 Y

q

E eθFq1q

(7.16)

since the variables (Mqx)q are independent. But E eθFq1q

= 1 +θE(Fq1q) +X

j≥2

θj

j!E(Fqj1q) andE(Fq1q) =E(Fq)−E(Fq1cq) =−E(Fq1cq).

Next note that if for any integerl, lN˜ <sup

t

Mqx|[t−A,t)≤(l+ 1) ˜N then

|Fq| ≤xb[(l+ 1)ηη+ 1] +xE(Z).

Hence, cutting Ωcq in slices of the type {lN˜ < suptMqx [t−A,t) ≤ (l+ 1) ˜N } and using Lemma3, we obtain by takingC3large enough,

|E(Fq1q)|=|E(Fq1cq)| ≤ same way, one can bound

E(Fqj1q)≤E(Fq21q)zbj−2,

using that ln(1 +u) ≤ u. It is sufficient now to recognize a step of the proof of the Bernstein inequality (weak version see [41, p25]). Since kz1 = α,η,sbT1−α0/(log(T)),

one can chooseα0>1, C1andC2in the definition (7.13) ofu(not depending onb) such thatu/2−kz1≥√

2kzvz+13zbzfor some z=C4log(T), whereC4is a constant. Hence

D≤exp

−θ(p

2kzvz+1

3zbz) +kX

j≥2

zvzbj−2θj j!

.

One can choose accordinglyθ (as for the proof of the Bernstein inequality) to obtain a bound ine−z. It remains to chooseC4large enough and only depending onα, η, Aandf0 to guarantee thatD≤e−zα,η,A,f0T−α.This concludes the proof of the proposition.

7.4.4. Proof of Proposition4

Let Q denote a measure such that under Q the distribution of the full point process restricted to (−∞,0] is identical to the distribution underPand such that on (0,∞) the process consists of independent components each being a homogeneous Poisson process with rate 1. Furthermore, the Poisson processes should be independent of the process on (−∞,0]. From Corollary 5.1.2 in [36] the likelihood process is given by

Lt= exp M t−X

m

Z t 0

λ(m)u du+X

m

Z t 0

logλ(m)u dNu(m)

!

and we have fort≥0 the relation

EPκt(f)2=EQκt(f)2Lt, (7.17) where EP and EQ denote the expectation with respect to P and Q respectively. Let, furthermore, ˜N1 =N[−1,0) denote the total number of points on [−1,0). Proposition4 will be an easy consequence of the following lemma.

Lemma 4. If the point process is stationary under P, if ed≤λ(m)t ≤a(N1+ ˜N1) +b

for t ∈[0,1] and for constants d∈ R and a, b > 0, and if EP(1 +ε)N˜1 < ∞ for some ε >0then for any f,

Q(f,f)≥ζ||f||2 (7.18)

for some constantζ >0.

Proof. We use H¨olders inequality on κ1(f)p2L

1 p

1 andκ1(f)2qL

1 p

1 to get EQκ1(f)2≤ EQκ1(f)2L11p

EQκ1(f)2L

q p

1

1q

=Q(f,f)1p

EQκ1(f)2L1−q1 1q

(7.19)

where 1p +1q = 1. We chooseq≥1 (and thusp) below to make q−1 sufficiently small.

For the left hand side we have by independence of the homogeneous Poisson processes that iff = (µ,(g`)`=1,...,M),

EQκ1(f)2 = (EQκ1(f))2+VQκ1(f)

= µ+X

`

Z 1 0

g`(u)du

!2

+X

`

Z 1 0

g`(u)2du.

Exactly as on page 32 in [52] there existsc0 >0 such that EQκ1(f)2≥c0 µ2+X

`

Z 1 0

g`2(u)du

!

=c0||f||2. (7.20) To bound the second factor on the right hand side in (7.19) we observe, by assumption, that we have the lower bound

L1≥eM(1−b)e(d−aM)N1e−aMN˜1

on the likelihood process. Under Qwe have that (κ1(f), N1) and ˜N1 are independent, and withρ=e(q−1)(aM−d) and ˜ρ=e(q−1)(aM) we get that

EQκ1(f)2L1−q1 ≤e(q−1)M(b−1)

EQρ˜N˜1EQκ1(f)2ρN1.

Here we chooseqsuch that ˜ρis sufficiently close to 1 to make sure thatEQρ˜N˜1 =EPρ˜N˜1 <

∞(see Proposition2). Moreover, by Cauchy-Schwarz’ inequality κ21(f)≤ µ2+X

`

Z 1−

0

g`2(1−u)dNu(`)

!

(1 +N1). (7.21) Under Q the point processes on (0,∞) are homogeneous Poisson processes with rate 1 and N1, the total number of points, is Poisson. This implies that conditionally on (N1(1), . . . , N1(M)) = (n(1), . . . , n(M)) the n(m)-points for them’th process are uniformly distributed on [0,1], hence

EQκ1(f)2L1−q1 ≤ µ2+X

`

Z 1 0

g`2(u)du

!

e(q−1)M(b−1)EQρ˜N˜1EQ(1 +N1)2ρN1

| {z }

c00

=c00||f||2. (7.22) Combining (7.20) and (7.22) with (7.19) we get that

c0||f||2≤(c00)1q||f||2qQ(f,f)1p or by rearranging that

Q(f,f)≥ζ||f||2 withζ= (c0)p/(c00)p−1.

For the Hawkes process it follows that ifν(m)>0 and if sup

t∈[0,1]

h(m)` (t)<∞

forl, m= 1, . . . , M then fort∈[0,1] we haveed≤λ(m)t ≤a(N1+ ˜N1) +b with d= logν(m), a= max

l sup

t∈[0,1]

h(m)` (t), b=ν(m).

Proposition2proves that there existsε >0 such thatEP(1 +ε)N˜1 <∞. This completes the proof of Proposition4.

Dans le document 1.3. Notation and overview of the paper (Page 43-52)

Documents relatifs