• Aucun résultat trouvé

Markov Random Geometric Graph (MRGG): A Growth Model for Temporal Dynamic Networks

N/A
N/A
Protected

Academic year: 2021

Partager "Markov Random Geometric Graph (MRGG): A Growth Model for Temporal Dynamic Networks"

Copied!
30
0
0

Texte intégral

(1)

HAL Id: hal-02865542

https://hal.archives-ouvertes.fr/hal-02865542v2

Preprint submitted on 12 Feb 2021

HAL is a multi-disciplinary open access archive for the deposit and dissemination of sci- entific research documents, whether they are pub- lished or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.

Markov Random Geometric Graph (MRGG): A Growth Model for Temporal Dynamic Networks

Yohann de Castro, Quentin Duchemin

To cite this version:

Yohann de Castro, Quentin Duchemin. Markov Random Geometric Graph (MRGG): A Growth Model for Temporal Dynamic Networks. 2021. �hal-02865542v2�

(2)

A Growth Model for Temporal Dynamic Networks

Yohann De Castro1 Quentin Duchemin2

Abstract

We introduce Markov Random Geometric Graphs (MRGGs), a growth model for temporal dynamic networks. It is based on a Markovian latent space dynamic: consecutive latent points are sampled on the Euclidean Sphere using an unknown Markov kernel; and two nodes are connected with a prob- ability depending on a unknown function of their latent geodesic distance.

More precisely, at each stamp-timekwe add a latent pointXksampled by jumping from the pre- vious oneXk−1in a direction chosen uniformly Ykand with a lengthrkdrawn from an unknown distribution called thelatitude function. The con- nection probabilities between each pair of nodes are equal to theenvelope functionof the distance between these two latent points. We provide theo- retical guarantees for the non-parametric estima- tion of the latitude and the envelope functions.

We propose an efficient algorithm that achieves those non-parametric estimation tasks based on an ad-hoc Hierarchical Agglomerative Cluster- ing approach. As a by product, we show how MRGGs can be used to detect dependence struc- ture in growing graphs and to solve link prediction problems.

1. Introduction

In Random Geometric Graphs (RGG), nodes are sampled in- dependently in latent spaceRd. Two nodes are connected if their distance is smaller than a threshold. A thorough prob- abilistic study of RGGs can be found in (Penrose,2003).

RGGs have been widely studied recently due to their abil- ity to provide a powerful modeling tool for networks with

1Institut Camille Jordan, ´Ecole Centrale de Lyon, Lyon, France

2LAMA, Univ Gustave Eiffel, Univ Paris Est Cr´eteil, CNRS F-77447 Marne-la-Vall´ee, France.. Correspondence to: ydc

<yohann.de-castro@ec-lyon.fr>, qd<quentin.duchemin@univ- eiffel.fr>.

spatial structure. We can mention applications in bioinfor- matics (Higham et al.,2008) or analysis of social media (Hoff et al.,2002). One main feature is to uncover hidden representation of nodes using latent space and to model interactions by relative positions between latent points.

Furthermore, nodes interactions may evolve with time. In some applications, this evolution is given by the arrival of new nodes as in online collection growth (Lo et al.,2017), online social network growth (Backstrom et al.,2006;Jin et al.,2001), or outbreak modeling (Ugander et al.,2012) for instance. The network is growing as more nodes are entering. Other time evolution modelings have been studied, we refer to (Rossetti and Cazabet,2018) for a review.

A natural extension of RGG consists in accounting this time evolution. In (D´ıaz et al.,2008), the expected length of connectivity and dis-connectivity periods of the Dynamic Random Geometric Graph is studied: each node choose at random an angle in [0,2π) and make a constant step size move in that direction. In (Staples,2009), a random walk model for RGG on the hypercube is studied where at each time step a vertex is either appended or deleted from the graph. Their model falls into the class of Geometric Markovian Random Graphs that are generally defined in (Clementi et al.,2009).

As far as we know, there is no extension of RGG to growth model for temporal dynamic networks. For the first time, in this paper, we consider a Markovian dynamic on the latent space where the new latent point is drawn with respect to the latest latent point and some Markov kernel to be estimated.

Estimation of graphon in RGGs: the Euclidean sphere case Random graphs with latent space can be defined us- ing agraphon, see (Lov´asz,2012). A graphon is a kernel function that defines edge distribution. In (Tang et al.,2013), Tang and al. prove that spectral method can recover the ma- trix formed by graphon evaluated at latent points up to an orthogonal transformation, assuming that graphon is a posi- tive definite kernel (PSD). Going further, algorithms have been designed to estimate graphons, as in (Klopp et al., 2017) which provide sharp rates for the Stochastic Block Model (SBM). Recently, the paper (Castro et al., 2020) provides a non-parametric algorithm to estimate RGGs on

(3)

Figure 1.Graphical model of the MRGG model: Markovian dy- namics on Euclidean sphere where we jump fromXkontoXk+1. TheYkencodes direction of jump whilerkencodes its distance, see (2).

Euclidean spheres, without PSD assumption.

We present here RGG on Euclidean sphere. Givennpoints X1, X2, . . . , Xnon the Euclidean sphereSd−1, we set an edge between nodes i and j (where i, j ∈ [n], i 6= j) with independent probabilityp(hXi, Xji). The unknown functionp: [−1,1]→[0,1]is called theenvelope function.

This RGG is a graphon model with a symmetric kernelW given byW(x, y) =p(hx, yi). Once the latent points are given, independently draw the random undirected adjacency matrixAby

Ai,j∼ B(p(hXi, Xji)), i < j

with Bernoulli r.v. drawn independently (set zero on the diagonal and complete by symmetry), and set

Tn:= 1

n(p(hXi, Xji))i,j∈[n] and Tˆn:= 1 nA, (1) We do not observe the latent point and we have to estimate the envelopepfromAonly. A standard strategy is to remark thatTˆnis a random perturbation ofTnand to dig intoTnto uncoverp.

One important feature of this model is that the interactions between nodes is depicted by a simple object: the enve- lope functionp. The envelope summarises how individuals connect each others given their latent positions. Standard examples (Bubeck et al.,2016) are given bypτ(t) =1{t≥τ}

where one connects two points as soon as their geodesic distance is below some threshold. The non-parametric es- timation ofpis given by (Castro et al.,2020) where the authors assume that latent pointsXiare independently and uniformly distributed on the sphere, which will not be the case in the present paper.

A new growth model: the latent Markovian dynamic Consider RGGs where latent points are sampled with Markovian jumps, the Graphical Model under consider- ation can be found in Figure 1. Namely, we samplen pointsX1, X2, . . . , Xnon the Euclidean sphereSd−1using

(a) Envelope function (b) Latitude function

Figure 2.Non-parametric estimation of envelope and latitude func- tions using algorithms of Sections2and3. We built a graph of 1500nodes sampled on the sphereS2 and using envelope and latitude (dot orange curves) defined in Section5by (10). The estimated envelope is thresholded to get a function in[0,1]and the estimated latitude function is normalized with integral1(plain blue lines).

a Markovian dynamic. We start by sampling randomlyX1 onSd−1. Then, for anyi∈ {2, . . . , n}, we sample

• a unit vectorYi∈Sd−1uniformly, orthogonal toXi−1.

• a realri∈[−1,1]encoding the distance betweenXi−1 and Xi, see (3). ri is sampled from a distribution fL : [−1,1]→[0,1], called thelatitude function.

thenXiis defined by

Xi =ri×Xi−1+ q

1−r2i ×Yi. (2) This dynamic can be pictured as follows. Consider that Xi−1 is the north pole, then chose uniformly a direction (i.e., a longitude) and, in a independent manner, randomly move along the latitudes (the longitude being fixed by the previous step). The geodesic distanceγidrawn on the lati- tudes satisfies

γi= arccos(ri), (3) where random variableri=hXi, Xi−1ihas densityfL(ri).

The resulting model will be referred to as the Markov Ran- dom Geometric Graph (MRGG) and is described with Fig- ure1.

Temporal Dynamic Networks: MRGG estimation strat- egy Seldom growth models exist for temporal dynamic network modeling, see (Rossetti and Cazabet,2018) for a review. In our model, we add one node at a time making a Markovian jump from the previous latent position. It results in

the observation of(Ai,j)1≤j≤i−1at timeT =i , as pictured in Figure1. Namely, we observe how a new node connects to the previous ones. For such dynamic, we aim at estimating the model, namely envelopepand

(4)

respectively latitudefL. These functions capture in a simple function onΩ = [−1,1]the range of interaction of nodes (represented byp) and respectively the dynamic of the jumps in latent space (represented byfL), where, in abscissaΩ, valuesr = hXi, Xji near 1 corresponds to close point Xi'Xjwhile values close to−1corresponds to antipodal pointsXi' −Xj. These functions may be non-parametric.

From snapshots of the graph at different time steps, can we recover envelope and latitude functions?This paper proves that it is possible under mild conditions on the Markovian dynamic of the latent points and our approach is summed up with Figure3.

Fundamental result Spectral convergence ofTˆnunder Markovian dynamic, see Section2.1

Guarantee for the recovery of: Algorithm (a)envelopep, see (5) ↔ SCCHEi (b)latent distancesri, see (9) ↔ HEiC2019 Figure 3.Presentation of our method to recover the envelope and the latitude functions.

Define λ(Tn) := (λ1, . . . , λn) and resp. λ( ˆTn) :=

(ˆλ1, . . . ,ˆλn) the spectrum of Tn and resp. Tˆn, see (1).

Building clusters fromλ( ˆTn), Algorithm1(SCCHEi) esti- mates the spectrum of envelopepwhile Algorithm3(Araya and De Castro,2019) (HEiC, see SectionEin Appendix) extractsdeigenvectors ofTˆnto uncover the Gram matrix of the latent positions. Both can then be used to estimate the unknown functions of our model (see Figure2).

Previous works Non-parametric estimation of RGGs on Euclidean sphere has been investigated in (Castro et al., 2020) with i.i.d. latent points. Estimation of latent point relative distances with HEiC Algorithm has been introduced in (Araya and De Castro,2019) under i.i.d. latent points assumption. Phase transitions on the detection of geome- try in RGGs (against Erd¨os R´enyi alternatives) has been investigated in (Bubeck et al.,2016).

For the first time, we introduce latitude function and non- parametric estimations of envelope and latitude using new results on kernel matrices concentration with dependent variables (see Appendix).

Outline Sections2and3present the estimation method with new theoretical results under Markovian dynamic.

These new results are random matrices operator norm con- trol and resp. U-statistics control under Markovian dynamic,

shown in the Appendix at SectionGand resp. SectionF. The envelopeadaptiveestimate is built from a size constrained clustering (Algorithm1) tuned by slope heuristic (6), and the latitude function estimate (see Section3.1) is derived from estimates of latent distancesri. Sections5and6inves- tigate synthetic data experiments1. We propose heuristics to solve link prediction problems and to test for a Marko- vian dynamic. Our method can handle random graphs with logarithmic growth node degree (i.e., new comer at time T =nconnects toO(logn)previous nodes), referred to as relatively sparsemodels, see Section4.

Notations Consider a dimensiond≥3. Denote byk · k2

(resp.h·,·i) the Euclidean norm (resp. inner product) onRd. Consider thed-dimensional sphereSd−1 := {x ∈ Rd : kxk2 = 1}and denote by σthe uniform distribution on Sd−1. For two real valued sequences(un)n∈Nand(vn)n∈N, denoteun =

n→∞O(vn)if there existk1 >0andn0 ∈N such that∀n > n0,|un| ≤k1|vn|.

Given two sequences x, y of reals–completing finite se- quences by zeros–such thatP

ix2i +yi2 <∞, we define the`2rearrangement distanceδ2(x, y)as

δ22(x, y) := inf

π∈S

X

i

(xi−yπ(i))2,

whereSis the set of permutations with finite support. This distance is useful to compare two spectra.

2. Nonparametric estimation of the envelope function

One can associate withW(x, y) = p(hx, yi)the integral operatorTW : L2(Sd−1) → L2(Sd−1)such that for any g∈L2(Sd−1),

∀x∈Sd−1, (TWg)(x) = Z

Sd−1

g(y)p(hx, yi)dσ(y), wheredσis the Lebesgue measure onSd−1. The operator TW is Hilbert-Schmidt and it has a countable number of bounded eigenvaluesλk with zero as only accumulation point. The eigenfunctions ofTW have the remarkable prop- erty that they do not depend onp(see (Dai,2013) Lemma 1.2.3): they are given by the real Spherical Harmonics. We denoteHlthe space of real Spherical Harmonics of degree lwith dimensiondland with orthonormal basis(Yl,j)j∈[dl] where

dl:= dim(Hl) =

1 ifl= 0

d ifl= 1

l+d−1 l

l+d−3l−2

otherwise.

1Our code is available at https://github.com/quentin- duchemin/Markovian-random-geometric-graph.git

(5)

We define also for allR ∈N,R˜ :=PR

l=0dl. We end up with the following spectral decomposition

p(hx, yi) =X

l≥0

pl X

1≤j≤dl

Yl,j(x)Yl,j(y) =X

k≥0

pkckGβk(hx, yi), (4) where λ(TW) = {p0, p1, . . . , p1, . . . , pl, . . . , pl, . . .} meaning that each eigenvaluepl has multiplicitydl; and Gβk is the Gegenbauer polynomial of degree k with pa- rameterβ := d−22 andck := 2k+d−2d−2 (see AppendixC).

Sincepis bounded, one hasp∈L2((−1,1), wβ)where the weight functionwβ is defined bywβ(t) := (1−t2)β−12, and it can be decomposed asp ≡ P

k≥pkckGβk and the Gegenbauer polynomialsGβk are an orthogonal basis of wβ(t) := (1−t2)β−12.

Weighted Sobolev space The spaceZws

β((−1,1))with regularity s > 0 is defined as the set of functionsg = P

k≥0gkckGβk ∈L2((−1,1), wβ)such that kgkZs

((−1,1)) :=

" X

l=0

dl|gl|2(1 + (l(l+ 2β))s)

#1/2

<∞.

2.1. Integral operator spectrum estimation with dependent variables

One key result is a new control ofU-statistics with latent Markov variables (see Section F) and it makes use of a Talagrand’s concentration inequality for Markov chains (see (Adamczak,2007)). This article follows the hypothesis made on the Markov chain(Xi)i≥1by (Adamczak,2007).

Namely, we referred to as “mild conditions” the assumption that the latitude function fL is bounded away from zero withkfLk<∞; and the first and the second regeneration times associated with thesplitchain to have sub-exponential tail. Those assumptions are fully described in sectionF.

Theorem1is a theoretical guarantee for a random matrix approximation of the spectrum of integral operator with dependentlatent variables. Theorem3in AppendixGgives explicitly the constants hidden in the big O below which depend on the spectral gap of the Markov chain(Xi)i≥1. Theorem 1 Assume mild conditions on the Markov chain (Xi)i≥1and assume the envelopephas regularitys >0.

Then, it holds Eh

δ22(λ(TW), λ(Tn))∨δ22(λ(TW), λRopt( ˆTn))i

= O

n log2(n)

2s+d−12s ! ,

with λRopt( ˆTn) = (ˆλ1, . . . ,λˆR˜opt,0,0, . . .)and Ropt = b n/log2(n)2s+d−11

cwhereˆλ1, . . . ,λˆn are the eigenval- ues ofTˆnsorted in decreasing order of magnitude.

Remark In Theorem1and Theorem2, note that we re- cover, up to alogfactor, theminimax rate of non-parametric estimationofs-regular functions on a space of (Riemannian) dimensiond−1. Even with i.i.d. latent variables, it is still an open question to know if this rate is the minimax rate of non-parametric estimation of RGGs.

Eq.(4) shows that one could use an approximation of (pk)k≥1to estimate the envelopepand Theorem1states we can recover(pk)k≥1up to a permutation. The problem of finding such a permutation is NP-hard and we introduce in the next section an efficient algorithm to fix this issue.

2.2. Size Constrained Clustering Algorithm

Note the spectrum ofTW is given by (pl)l≥0 where pl has multiplicity dl. In order to recover envelope p, we build clusters from eigenvalues ofTˆnwhile respecting the dimensiondlof each eigen-space ofTW. In (Castro et al., 2020), an algorithm is proposed testing all permutations of {0, . . . , R}for a given maximal resolutionR. To bypass the high computational cost of such approach, we propose an efficient method based on the tree built fromHierarchical Agglomerative Clustering(HAC).

In the following, for anyν1, . . . , νn ∈ R, we denote by HAC({ν1, , . . . , νn}, dc)the tree built by a HAC on the real valuesν1, . . . , νn using the complete linkage functiondc defined by∀A, B⊂R,dc(A, B) = maxa∈Amaxb∈Bka−

bk2. Algorithm1describes our approach.

Algorithm 1Size Constrained Clustering for Harmonic Eigenvalues (SCCHEi).

Data: ResolutionR, matrixTˆn=n1A, dimensions(dk)Rk=0. 1: Letλˆ1, . . . ,λˆnbe the eigenvalues ofTˆnsorted in decreasing

order of magnitude.

2: SetP:={ˆλ1, . . . ,ˆλR˜}anddims= [d0, d1, . . . , dR].

3: whileAll eigenvalues inPare not clustereddo 4: tree←HAC(nonclustered eigenvalues inP,dc) 5: for d∈dims do

6: Search for a cluster of sizedintreeas close as possible to the root.

7: if such a cluster Cd exists then

Update(dims, tree,Cd, d).

8: end for

9: ford∈dimsdo

10: Search for the groupCintreewith a size larger thand and as close as possible tod.

11: if such a group exists then Update(dims, tree,C, d) elseGo to line3.

12: end for 13: end while

Return:Cd0, . . . ,CdR,{λˆR+1˜ , . . . ,λˆn}

(6)

Algorithm 2Update(dims, tree,C, d).

1: Save the subsetCdconsisting of thedeigenvalues inC with the largest absolute values.

2: Delete fromtreeall occurrences to eigenvalues inCd and deletedfromdims.

2.3. Adaptation: Slope heuristic as model selection of Resolution

A data-driven choice of model sizeRcan be done byslope heuristic, see (Arlot,2019) for a nice review. One main idea of slope heuristic is to penalize the empirical risk by κpen( ˜R) and to calibrate κ > 0. If the sequence (pen( ˜R))R˜ is equivalent to the sequence of variances of the population risk of empirical risk minimizer (ERM) as model sizeR˜grows, then, penalizing the empirical risk (as done in (6)), one may ultimately uncover an empirical ver- sion of theU-shaped curve of the population risk. Hence, minimizing it, one builds a model sizeRˆthat balances be- tween bias (under-fittingregime) and variance (over-fitting regime). First, note that empirical risk is given by the intra- class variance below.

Definition 1 For any output(Cd0, . . . ,CdR,Λ) of the Al- gorithm SCCHEi, the thresholded intra-class variance is defined by

IR:= 1 n

R

X

k=0

X

λ∈Cdk

λ− 1 dk

X

λ0∈Cdk

λ0

2

+X

λ∈Λ

λ2

, and the estimations(ˆpk)k≥0of the eigenvalues(pk)k≥0is given by

∀k∈N, pˆk = ( 1

dk

P

λ∈Cdkλ ifk∈ {0, . . . ,R}ˆ

0 otherwise.

(5) Second, as underlined in the proof of Theorem1(see The- orem3in the Appendix), the estimator’s variance of our estimator scales linearly inR.˜

Hence, we apply Algorithm SCCHEi forRvarying from 0 toRmax := max{R ≥ 0 : R˜ ≤ n}to compute the thresholded intra-class varianceIR(see Definition1) and given someκ >0, we select

R(κ)∈ arg min

R∈{0,...,Rmax}

nIR+κ R˜ n o

. (6)

The hyper-parameterκcontrolling the bias-variance trade- off is set to2κ0whereκ0is the value ofκ >0leading to the “largest jump” of the functionκ7→R(κ). OnceRˆ:=

R(2κ0)has been computed, we approximate the envelope functionpusing (5) (see (19) in Appendix for the closed

form). In AppendixD, we describe this slope heuristic on a concrete example and our results can be reproduced using the notebook in the Supplementary Material.

3. Nonparametric estimation of the latitude function

3.1. Our approach to estimate the latitude function in a nutshell

In Theorem 2(see below), we show that we are able to estimate consistently the pairwise distances encoded by the Gram matrixGwhere

G:= 1

n(hXi, Xji)i,j∈[n].

Taking the diagonal just above the main diagonal (referred to assuperdiagonal) ofGˆ- an estimate of the matrixGto be specified - we get estimates of the i.i.d. random vari- ables(hXi, Xi−1i)2≤i≤n = (ri)2≤i≤n sampled from fL. Using (ˆri)2≤i≤n the superdiagonal ofnG, we can buildˆ a kernel density estimator of the latitude functionfL. In the following, we describe the algorithm used to build our estimatorGˆwith theoretical guarantees.

3.2. Spectral gap condition and Gram matrix estimation

The Gegenbauer polynomial of degree one is defined by Gβ1(t) = 2βt, ∀t ∈ [−1,1]. As a consequence, us- ing theaddition theorem(see (Dai,2013, Lem.1.2.3 and Thm.1.2.6)), the Gram matrixGis related to the Gegen- bauer polynomial of degree one. More precisely, for any i, j∈[n]it holds

Gi,j = 1

2βnGβ1(hXi, Xji) = 1 nd

d

X

k=1

Y1,k(Xi)Y1,k(Xj).

(7) Denoting V ∈ Rn×d the matrix with columnsvk :=

1

n(Y1,k(X1), . . . , Y1,k(Xn))fork∈[d], (7) becomes G:=1

dV(V)>.

We will prove that fornlarge enough there exists a matrix Vˆ ∈ Rn×d where each column is an eigenvector ofTˆn, such thatGˆ:= 1dVˆVˆ>approximatesGwell, in the sense that the normkG−Gkˆ Fconverges to0. To choose thed eigenvectors of the matrixTˆnthat we will use to build the matrixVˆ, we need the following spectral gap condition

:= min

k∈N, k6=1|p1−pk|>0. (8) This condition will allow us to apply Davis-Kahan type inequalities.

(7)

Now, thanks to Theorem1, we know that the spectrum of the matrixTˆnconverges towards the spectrum of the inte- gral operatorTW. Then, based on (7), one can naturally think that extracting thedeigenvectors of the matrixTˆn

related with the eigenvalues that converge towardsp1, we can approximate the Gram matrixGof the latent positions.

Theorem2proves that the latter intuition is true with high probability under the spectral gap condition (8). The algo- rithm HEiC (Araya and De Castro,2019) (See SectionEfor a presentation) aims at identifying the above mentionedd eigenvectors of the matrixTˆnto build our estimate of the Gram matrixG.

Theorem 2 Assume mild conditions on the Markov chain (Xi)i≥1, assume∆>0, and assume that graphonW has regularitys >0. We denoteVˆ ∈Rn×dthedeigenvectors of the matrixTˆnassociated with the eigenvalues returned by the algorithm HEiC and we defineGˆ := 1dVˆVˆ>. Then fornlarge enough and for some constantD >0, it holds with probability at least1−5/n2,

kG−Gkˆ F ≤D n

log2(n)

2s+d−1−s

. (9)

Based on Theorem2, we propose a kernel density approach to estimate the latitude function fL based on the super- diagonal of the matrixG, namelyˆ

ˆ

ri:=Gbi−1,i

i∈{2,...,n}. In the following, we denotefˆLthis estimator.

4. Relatively Sparse Regime

Although this paper deals with the so-calleddenseregime (i.e. when the expected number of neighbors of each node scales linearly withn), our results may be generalized to the relatively sparsemodel connecting nodesiandjwith prob- abilityW(Xi, Xj) =ζnp(hXi, Xji)whereζn∈(0,1]sat- isfieslim inf

n ζnn/logn≥Z for some universal constant Z >0.

In the relatively sparse model, one can show following the proof of Theorem1that the resolution should be cho- sen as Rˆ =

n

1+ζnlog2n

2s+d−11

. Specifying that λ = (p0, p1, . . . , p1, p2, . . .) and Tˆn = A/n, Theorem1 be-

comes for a graphon with regularitys >0

E

"

δ22 λ,λ( ˆTn) ζn

!#

=O

n

1 +ζnlog2n

2s+d−1−2s ! .

Figure4illustrates the estimation of the latitude and the envelope functions in some relatively sparse regimes.

Figure 4.Results of our algorithms for graph of size2,000with functions of Eq.(10) and sparsity parameterζn= logkn/n,k∈ {2,3,4}.

5. Experiments

We test our method usingd= 3and considering the follow- ing functions

p:x7→1x≥0, and fL:x7→

1

2g(1−x; 2,2) ifx≥0

1

2g(1 +x; 2,2) otherwise , (10) whereg(·; 2,2)is the density of the beta distribution with pa- rameters(2,2). Figure5.(a)presents theδ2-error between the spectra of the true envelope (resp. latitude) function and the estimated one when the size of the graph is increasing.

In Figure5.(b), we propose a visualization of the cluster- ing performed by SCCHEi withn = 1000and R = 4.

Blue crosses represent theR˜ eigenvalues ofTˆn with the largest magnitude, which are used to form clusters corre- sponding to the five-first spherical harmonic spaces. The red circles are the estimated eigenvalues(ˆpk)0≤k≤4(plot- ted with multiplicity) defined from the clustering given by our algorithm SCCHEi (see (5)). Those results show that SCCHEi achieves a relevant clustering of the eigenvalues ofTˆnwhich allows us to recover the envelope function (see Figure2.(a)).

(a)δ2error on envelope and lat- itude functions.

(b) Clustering adjacency eigen- values (R= 4).

Figure 5.Non-parametric estimation of envelope and latitude us- ing algorithms described in Sections2and3. In (a), bars represent standard deviation ofδ2errors between true and estimated func- tions.

6. Applications

In this section, we apply the MRGG model to link prediction and hypothesis testing in order to demonstrate the usefulness

(8)

of our approach as well as the estimation procedure.

6.1. Markovian Dynamic Testing

As a first application of our model, we propose a hypothesis test to statistically distinguish between an independent sam- pling the latent positions and a Markovian dynamic. The null is then set toH0:nodes are independent and uniformly distributed on the sphere(i.e.,no Markovian dynamic). Our test is based on estimatefˆL of latitude and thus the null can be rephrased asH0: fL =fL0wherefL0is the latitude of uniform law, dynamic is then i.i.d. dynamic. Figure6 shows the power of a hypothesis test with level5%(Type I error). One can use anyblack-box goodness-of-fit testcom- paringfˆLtofL0, and we chooseχ2-test discretizing(−1,1) in70regular intervals. Rejection region is calibrated (i.e.,

Figure 6.Hypothesis testing.

threshold of theχ2-test here) byMonte Carlo simulations under the null. It allows us to control Type I error as de- picted by dotted blue line. We choose alternative given by Heaviside envelope and latitudefLof (10). We run our al- gorithm to estimate latitude from which we sample a batch to compute theχ2-test statistic. We see that for graphs of size larger than1,000, the rejection rate is almost1under the alternative (Type II error is almost zero), the test is very powerful.

6.2. Link Prediction

Suppose that we observe a graph withnnodes. Link pre- diction is the task that consists in estimating the probability of connection between a given node of the graph and the upcoming node.

6.2.1. BAYESLINKPREDICTION

We propose to show the usefulness of our model solving a link prediction problem. Let us recall that we do not es- timate the latent positions but only thepairwise distances (embedding task is not necessary for our purpose). De-

noting projX

n(·) the orthogonal projection onto the or- thogonal complement ofSpan(Xn), the decomposition of hXi, Xn+1idefined by

hXi, XnihXn, Xn+1i+p

1− hXn, Xn+1i2

×p

1− hXi, Xni2h projX n(Xi) kprojX

n(Xi)k2

, Yn+1i, (11) shows that latent distances are enough for link prediction.

Indeed, it can be achieved using a forward step on our Markovian dynamic, giving the posterior probability (see Definition2)ηi(D1:n)defined by

Z

[−1,1]2

p

hXi, Xnir+p

1−r2p

1− hXi, Xni2u

fL(r)drdu 2 . (12)

Definition 2 (Posterior probability function)

The posterior probability function η is defined for any latent pairwise distances D1:n = (hXi, Xji)1≤i,j≤n ∈ [−1,1]n×nby

∀i∈[n], ηi(D1:n) =P(Ai,n+1= 1|D1:n), whereAi,n+1 ∼ B(p(hXi, Xn+1i)is a random variable that equals1if there is an edge between nodesiandn+ 1, and is zero otherwise.

We consider a classifierg(see Definition3) and an algorithm that, given some latent pairwise distancesD1:n, estimates Ai,n+1by putting an edge between nodesXiandXn+1if gi(D1:n)is1.

Definition 3 A classifier is a function which associates to any pairwise distancesD1:n= (hXi, Xji)1≤i,j≤n, a label (gi(D1:n))i∈[n] ∈ {0,1}n.

The risk of this algorithm is as inbinary classification, R(g,D1:n) := 1

n

n

X

i=1

P(gi(D1:n)6=Ai,n+1|D1:n)

= 1 n

n

X

i=1

(1−ηi(D1:n))1gi(D1:n)=1i(D1:n)1gi(D1:n)=0 ,

(13) where we used the independence between Ai,n+1 and gi(D1:n)conditionally onσ(D1:n). Pushing further this analogy, we can define the classification error of some clas- sifier gbyL(g) = E[R(g,D1:n)]. Proposition 1shows that the Bayes estimator - introduced in Definition4- is optimal for the risk defined in (13).

Definition 4 (Bayes estimator)

We keep the notations of Definition2. The Bayes estimator gof(Ai,n+1)1≤i≤nis defined by

∀i∈[n], gi(D1:n) =

1 ifηi(D1:n)≥ 12 0 otherwise.

(9)

Proposition 1 (Optimality of the Bayes classifier for the riskR)

We keep the notations of Definitions2and4. For any classi- fierg, it holds for alli∈[n],

P(gi(D1:n)6=Ai,n+1|D1:n)

−P(gi(D1:n)6=Ai,n+1|D1:n)

= 2

ηi(D1:n)−1 2

×E

1gi(D1:n)6=gi(D1:n)|D1:n , which immediately implies that

R(g,D1:n)≥ R(g,D1:n)and thereforeL(g)≥L(g).

6.2.2. HEURISTIC FORLINKPREDICTION

One natural method to approximate the Bayes classifier from the previous section is to use theplug-in approach. This leads to the MRGG classifier introduced in Definition5.

Definition 5 (The MRGG classifier)

For anynand anyi∈[n], we defineηˆi(D1:n)as Z

ˆ p

bri,nr+p

1−r2q

1−rb2i,nu

L(r)drdu

2 , (14)

wherepˆandfˆLdenote respectively the estimate of the enve- lope function and the latitude function with our method and wherebr:=nG. The MRGG classifier is defined byb

∀i∈[n], giM RGG(D1:n) =

1 ifηˆi(D1:n)≥ 12 0 otherwise.

To illustrate our approach we work with a graph of2,000 nodes with d = 3, Heaviside envelope and fL(r) =

1

2f(5,1)(r+12 )wheref(5,1)is the pdf of the Beta distribution with parameter(5,1). Figure7shows that we are able to recover the probabilities of connection of the nodes already present in the graph with the coming nodeXn+1. Using the decomposition ofhXi, Xn+1igiven by (11), orange crosses are computed using (12). Green plus are computed similarly replacingpandfLby their estimationspˆandfˆLfollowing (14). Blue stars are computed using (12) by replacingfLby

wβ

kwβk1 = 12 (sinced= 3) which amounts to consider that the points are sampled uniformly on the sphere.

In Figure8, we compare the risk of therandomclassifier - whose guessgi(D1:n)is a Bernoulli random variable with parameter given by the ratio of edges compared to complete graph - with the risk of the MRGG classifier (see Defini- tion5). Figure8shows that the risk of the MRGG classifier is significantly smaller than the risk of random classifier, it is clearly better than random. Moreover, Figure8shows that the MRGG classifier gives similar results compared to the optimal Bayes classifier.

Figure 7.Link predictions between the future nodeXn+1and the 10first nodesX1, . . . , X10.

Figure 8.Comparison between the risk (defined in (13)) of the MRGG classifier, the random classifier and the risk of the optimal Bayes classifier.

References

R. Adamczak. A tail inequality for suprema of unbounded empirical processes with applications to markov chains.

Electronic Journal of Probability, 13, 10 2007. doi: 10.

1214/EJP.v13-521.

E. Araya and Y. De Castro. Latent Distance Estimation for Random Geometric Graphs. InAdvances in Neural Information Processing Systems, pages 8721–8731, 2019.

S. Arlot. Minimal penalties and the slope heuristics: a survey. Journal de la Soci´et´e Franc¸aise de Statistique, 160(3):1–106, 2019.

L. Backstrom, D. Huttenlocher, J. Kleinberg, and X. Lan.

Group formation in large social networks: membership, growth, and evolution. InProceedings of the 12th ACM SIGKDD international conference on Knowledge discov- ery and data mining, pages 44–54, 2006.

A. S. Bandeira and R. van Handel. Sharp nonasymptotic bounds on the norm of random matrices with independent

(10)

entries. Ann. Probab., 44(4):2479–2506, 07 2016. doi:

10.1214/15-AOP1025. URLhttps://doi.org/10.

1214/15-AOP1025.

R. Bhatia. Matrix Analysis. Graduate Texts in Mathemat- ics. Springer New York, 1996. ISBN 9780387948461.

URL https://books.google.fr/books?id=

F4hRy1F1M6QC.

S. Bubeck, J. Ding, R. Eldan, and M. Z. R´acz. Testing for high-dimensional geometry in random graphs. Random Structures & Algorithms, 49(3):503–532, Jan 2016. ISSN 1042-9832. doi: 10.1002/rsa.20633. URLhttp://dx.

doi.org/10.1002/rsa.20633.

Y. D. Castro, C. Lacour, and T. M. P. Ngoc. Adaptive Estima- tion of Nonparametric Geometric Graphs. Mathematical Statistics and Learning, 2020.

A. E. Clementi, F. Pasquale, A. Monti, and R. Silvestri.

Information spreading in stationary Markovian evolving graphs. 2009 IEEE International Symposium on Parallel

& Distributed Processing, May 2009. doi: 10.1109/

ipdps.2009.5160986. URL http://dx.doi.org/

10.1109/IPDPS.2009.5160986.

Y. Dai, F.and Xu. Approximation theory and harmonic analysis on spheres and balls. Springer, 2013.

C. Dorea and A. Pereira. A note on a variation of Doeblin’s condition for uniform ergodicity of Markov chains.Acta Mathematica Hungarica, 110:287–292, 01 2006. doi:

10.1007/s10474-006-0023-y.

Q. Duchemin, Y. de Castro, and C. Lacour. Concentra- tion inequality for u-statistics of order two for uniformly ergodic markov chains, and applications, 2020.

J. D´ıaz, D. Mitsche, and X. P´erez-Gim´enez. On the connec- tivity of dynamic random geometric graphs, 01 2008.

J. Fan, B. Jiang, and Q. Sun. Hoeffding’s lemma for Markov Chains and its applications to statistical learning, 2018.

D. Ferr´e, L. Herv´e, and J. Ledoux. Limit theorems for stationary Markov processes with L2-spectral gap. An- nales de l’Institut Henri Poincar´e, Probabilit´es et Statis- tiques, 48(2):396–423, May 2012. ISSN 0246-0203. doi:

10.1214/11-aihp413. URL http://dx.doi.org/

10.1214/11-AIHP413.

P. Gilles. The volume of convex bodies and Banach space geometry. Cambridge University Press, 1989.

D. J. Higham, M. Raˇsajski, and N. Prˇzulj. Fitting a ge- ometric graph to a protein–protein interaction network.

Bioinformatics, 24(8):1093–1099, 03 2008. ISSN 1367- 4803. doi: 10.1093/bioinformatics/btn079.

P. D. Hoff, A. E. Raftery, and M. S. Handcock. Latent Space Approaches to Social Network Analysis.Journal of the American Statistical Association, 97(460):1090–1098, 2002. doi: 10.1198/016214502388618906. URLhttps:

//doi.org/10.1198/016214502388618906.

B. Jiang, Q. Sun, and J. Fan. Bernstein’s inequality for general Markov Chains, 2018.

E. M. Jin, M. Girvan, and M. E. Newman. Structure of grow- ing social networks. Physical review E, 64(4):046132, 2001.

O. Klopp, A. B. Tsybakov, and N. Verzelen. Oracle in- equalities for network models and sparse graphon esti- mation. The Annals of Statistics, 45(1):316–354, Feb 2017. ISSN 0090-5364. doi: 10.1214/16-aos1454. URL http://dx.doi.org/10.1214/16-AOS1454.

V. Koltchinskii and E. Gin´e. Random Matrix Approximation of Spectra of Integral Operators. Bernoulli, 6, 02 2000.

doi: 10.2307/3318636.

C. Lo, J. Cheng, and J. Leskovec. Understanding Online Collection Growth Over Time: A Case Study of Pin- terest. In Proceedings of the 26th International Con- ference on World Wide Web Companion, WWW ’17 Companion, page 545–554, Republic and Canton of Geneva, CHE, 2017. International World Wide Web Con- ferences Steering Committee. ISBN 9781450349147.

doi: 10.1145/3041021.3054189. URLhttps://doi.

org/10.1145/3041021.3054189.

L. Lov´asz. Large Networks and Graph Limits. American Mathematical Society, 01 2012.

S. Meyn and R. Tweedie. Markov Chains and Stochastic Stability, volume 92. 01 1993. doi: 10.2307/2965732.

M. Penrose. Random Geometric Graphs, 01 2003.

G. O. Roberts and J. S. Rosenthal. General state space Markov chains and MCMC algorithms. Probability Sur- veys, 1(0):20–71, 2004. ISSN 1549-5787. doi: 10.1214/

154957804100000024. URLhttp://dx.doi.org/

10.1214/154957804100000024.

G. Rossetti and R. Cazabet. Community discovery in dy- namic networks: a survey. ACM Computing Surveys (CSUR), 51(2):1–37, 2018.

R. S. . G. Staples. Dynamic Geometric Graph Processes : Adjacency Operator Approach, 2009.

M. Tang, D. L. Sussman, and C. E. Priebe. Universally con- sistent vertex classification for latent positions graphs.

The Annals of Statistics, 41(3):1406–1430, Jun 2013.

ISSN 0090-5364. doi: 10.1214/13-aos1112. URL http://dx.doi.org/10.1214/13-AOS1112.

(11)

Supplementary Material

Markov Random Geometric Graph (MRGG):

A Growth Model for Temporal Dynamic Networks

Guidelines for the supplementary material SectionsAtoC: Basic definitions and Complements

In SectionAwe recall basic definitions on Markov chains which are required for SectionBwhere we describe some properties verified by the Markov chain(Xi)i≥1. SectionCprovides complementary results on the Harmonic Analysis on Sd−1which will be useful for our proofs.

SectionsDtoE: Algorithms and Experiments2

SectionDdescribes precisely the slope heuristic used to perform the adaptive selection of the model dimensionR. Section˜ E provides a complete description of the HEiC alogorithm used to extractd-eigenvectors of the adjacency matrix that will be used to estimate the Gram matrix of the latent positions.

SectionsFtoH: Proofs of theoretical results

Thereafter, we dig into the most theoretical part of the supplementary material. In SectionF, we provide a full description of the assumptions we made on the Markov chain(Xi)i≥1and which we have been referring to so far bymild conditions.

SectionFis also dedicated to the presentation of a concentration result for a particular U-statistic of the Markov chain (Xi)i≥1 that is an essential element of the proof of Theorem1which is provided in SectionG. Finally, the proof of

Theorem2can be found in SectionH.

A. Definitions for general Markov chains

We consider a state spaceEand a sigma-algebraΣonEwhich is a standard Borel space. We denote by(Xi)i≥1a time homogeneous Markov chain on the state space(E,Σ)with transition kernelP.

A.1. Ergodic and reversible Markov chains

Definition 6 (Roberts and Rosenthal,2004, section 3.2) (φ-irreducible Markov chains)

The Markov chain(Xi)i≥1is saidφ-irreducible if there exists a non-zeroσ-finite measureφonEsuch that for allA∈Σ withφ(A)>0, and for allx∈E, there exists a positive integern=n(x, A)such thatPn(x, A)>0(wherePn(x,·) denotes the distribution ofXn+1conditioned onX1=x).

Definition 7 (Roberts and Rosenthal,2004, section 3.2) (Aperiodic Markov chains)

The Markov chain (Xi)i≥1 with invariant distributionπis aperiodic if there do not existm ≥ 2 and disjoint subsets A1, . . . , Am ⊂EwithP(x, Ai+1) = 1for allx∈Ai (1 ≤i≤m−1), andP(x, A1) = 1for allx∈Am, such that π(A1)>0(and henceπ(Ai)>0for alli).

Definition 8 (Roberts and Rosenthal,2004, section 3.4) (Geometric ergodicity)

The Markov chain (Xi)i≥1 is said geometrically ergodic if there exists an invariant distribution π, ρ ∈ (0,1) and C:E→[1,∞)such that

kPn(x,·)−πkT V ≤C(x)ρn, ∀n≥0, π−a.ex∈E, wherekµkT V := supA∈Σ|µ(A)|.

Definition 9 (Roberts and Rosenthal,2004, section 3.3) (Uniform ergodicity)

The Markov chain(Xi)i≥1is said uniformly ergodic if there exists an invariant distributionπand constants0< ρ <1and

2Our code is available athttps://github.com/quentin-duchemin/Markovian-random-geometric-graph.git

(12)

L >0such that

kPn(x,·)−πkT V ≤Lρn, ∀n≥0, π−a.ex∈E, wherekµkT V := supA∈Σ|µ(A)|.

Remark:A Markov chain geometrically or uniformly ergodic admits a unique invariant distribution and is aperiodic.

Definition 10 A Markov chain is said reversible if there exists a distributionπsatisfying π(dx)P(x, dy) =π(dy)P(y, dx).

A.2. Spectral gap

This section is largely inspired from (Fan et al.,2018). Let us consider that the Markov chain(Xi)i≥1admits a unique invariant distributionπonSd−1.

For any real-valued,Σ-measurable functionh:E→R, we defineπ(h) :=R

h(x)π(dx). The set L2(E,Σ, π) :={h:π(h2)<∞}

is a Hilbert space endowed with the inner product hh1, h2iπ=

Z

h1(x)h2(x)π(dx), ∀h1, h2∈ L2(E,Σ, π).

The map

k · kπ:h∈ L2(E,Σ, π)7→ khkπ=p hh, hiπ,

is a norm onL2(E,Σ, π).k · kπnaturally allows to define the norm of a linear operatorTonL2(E,Σ, π)as Nπ(T) = sup{kT hkπ :khkπ= 1}.

To each transition probability kernelP(x, B)withx∈EandB ∈Σinvariant with respect toπ, we can associate a bounded linear operatorh7→R

h(y)P(·, dy)onL2(E,Σ, π). Denoting this operatorP, we get P h(x) =

Z

h(y)P(x, dy), ∀x∈E, ∀h∈ L2(E,Σ, π).

LetL02(π) :={h∈ L2(E,Σ, π) : π(h) = 0}. We define the absolute spectral gap of a Markov operator.

Definition 11 (Spectral gap) A Markov operatorP reversible admits a spectral gap1−λif λ:= sup

kP hkπ khkπ

: h∈ L02(π), h6= 0

<1.

The next result provides a connection between spectral gap and geometric ergodicity for reversible Markov chains.

Proposition 2 (Ferr´e et al.,2012, section 2.3)

A uniformly ergodic Markov chain admits a spectral gap.

(13)

B. Properties of the Markov chain

Let P be the Markov operator of the Markov chain(Xi)i≥1. By abuse of notation, we will also denoteP(x,·) the density of the measure P(x, dz)with respect todσ, the uniform measure on Sd−1. For any x, z ∈ Sd−1, we denote Rzx∈Rd×da rotation matrix sendingxtoz(i.e.Rzxx=z) and keepingSpan(x, z)fixed. In the following, we denote ed:= (0,0, . . . ,0,1)∈Rd.

B.1. Invariant distribution and reversibility for the Markov chain Reversibility of the Markov chain(Xi)i≥1.

Lemma 1 For allx, z∈Sd−1,P(x, z) =P(z, x) =P(ed, Rzedx).

Proof of Lemma1.

Using our model described in section2, we getX2=rX1+√

1−r2Y where conditionally onX1,Y is uniformly sampled onS(X1) := {q ∈ Sd−1 : hq, X1i = 0},and whererhas densityfL on[−1,1]. Let us consider a gaussian vector W ∼ N(0, Id). Using the Cochran’s theorem and Lemma2, we know that conditionally onX1, the random variable

W−hW,X1iX1

kW−hW,X1iX1k2 is distributed uniformly onS(X1).

Lemma 2 LetW ∼ N(0, Id). Then, kWWk

2 is distributed uniformly on the sphereSd−1. In the following, we denote=L the equality in distribution sense. We have conditionally onX1

ReXd

1

W − hW, X1iX1

kW− hW, X1iX1k2 =

Wˆ − hW , eˆ died

kWˆ − hW , eˆ diedk2

,

whereWˆ =ReXd

1W ∼ N(0, Id). Using Cochran’s theorem, we know thatWˆ − hW , eˆ diedis a centered normal vector with covariance matrix the orthographic projection matrix onto the spaceSpan(ed), leading to

Wˆ − hW , eˆ died

=L

Y 0

,

whereY ∼ N(0, Id−1). Using Lemma2, we conclude that conditionally onX1, the random variablekWW−hW,X−hW,X1iX1

1iX1k2 is distributed uniformly onS(X1)(because the distribution ofY is invariant by rotation).

We deduce that X2

=L rX1+p

1−r2 W− hW, X1iX1

kW − hW, X1iX1k2

=L rX1+p

1−r2 RXX1

2W0− hRXX1

2W0, X1iX1 kRXX1

2W0− hRXX1

2W0, X1iX1k2, whereW0:=RXX2

1W. Note thatW0∈Rdis also a standard centered gaussian vector because this distribution is invariant by rotation. SincehRXX1

2W0, X1i=hW0, X2iandkRXX1

2qk2=kqk2, ∀q∈Sd−1, we deduce that X2−rX1=L RXX1

2

p

1−r2 W0− hW0, X2iX2 kW0− hW0, X2iX2k2

. (15)

RXX2

1 is the rotation that sendsX1toX2keeping the other dimensions fixed. Let us denotea1:=X1,a2:= kXX2−rX1

2−rX1k2 and complete the linearly independent family(a1, a2)in an orthonormal basis ofRdgiven bya:= (a1, a2, . . . , ad). Then, the matrix ofRXX2

1 in the basisais

r −√

1−r2 0>d−2

√1−r2 r 0>d−2 0d−2 0d−2 Id−2

.

(14)

We deduce that

RXX1

2

−1

(X2−rX1) =RXX2

1(X2−rX1)

=kX2−rX1k2RXX2

1

X2−rX1 kX2−rX1k2

=kX2−rX1k2RXX2

1a2

=kX2−rX1k2h

−p

1−r2a1+ra2i

=−p

1−r2kX2−rX1k2X1+rX2−r2X1

=−(1−r2)X1+rX2−r2X1

=−X1+rX2. Going back to (15), we deduce that

X1

=L rX2+p 1−r2

W˜ − hW , X˜ 2iX2

kW˜ − hW , X˜ 2iX2k2

, (16)

whereW˜ =−W0is also a standard centered gaussian vector inRd. Thus, we proved the first equality of Lemma1. Based on (16) we have,

ReXd

2X1=L rReXd

2X2+p

1−r2RXed

2

W˜ − hW , X˜ 2iReXd

2X2 kW˜ − hW , X˜ 2iX2k2

=red+p

1−r2 RXed

2

W˜ − hRXed

2

W , e˜ died

kReXd

2

W˜ − hReXd

2

W , e˜ diedk2

, which proves thatP(ed, Rexd2x1) =P(x2, x1)for anyx1, x1∈Sd−1becauseReXd

2

W˜ is again a standard centered gaussian vector inRd.

Invariant distribution of the Markov chain.

Proposition 3 The uniform distribution on the sphereSd−1is an invariant distribution of the Marokv chain(Xi)i≥1. Proof of Proposition3.

Let us considerz∈Sd−1. We denotedσ≡dσdthe Lebesgue measure onSd−1anddσd−1the Lebesgue measure onSd−2. Using (Dai,2013, Section 1.1), it holdsbd:=R

Sd−1dσ= Γ(d/2)d/2. We have Z

x∈Sd−1

P(x, z)dσ(x) bd

= Z

x∈Sd−1

P(ed, Rezdx)dσ(x) bd

(Using Lemma1)

= Z

x∈Sd−1

P(ed, x)dσ(x)

bd (Using the change of variablex7→Rzedx)

= Z π

0

Z

Sd−2

P

ed, ξsinθ

cosθ

(sinθ)d−2dθdσd−1(ξ)

bd (Using (Dai,2013, Eq.(1.5.4) Section 1.5))

= Z 1

−1

Z

Sd−2

P

ed, ξ√

1−r2 r

(1−r2)d−32 drdσd−1(ξ) bd

= Z 1

−1

fL(r)dr× Z

Sd−2

1 bd−1

d−1 bd = 1

bd,

which proves that the uniform distribution on the sphere is an invariant distribution of the Markov chain.

Références

Documents relatifs

The agglomerates were constructed by extraction of spherical samples from a gran- ular bed prepared by compaction for different values of particle size span, and subjected

Accessorial fees Fees incurred by shippers and paid to carriers for performing freight services beyond the normal pickup and delivery, including excess waiting

Pour supprimer ce possible biais dans nos résultats et pouvoir observer le seul impact de notre proposition d’apprentissage de transfert, nous avons tout simplement décidé de

Although engineers still have a significantly higher hazard rate than all forms of PhDs prior to the reforms in 2006, and despite the fact that the coefficients for all types of PhD

And if one’s view is that something plays the role of S’s motivating reason just in case (a) S takes it to speak in favor of Φing, (b) is motivated by it to Φ, (c) is motivated by

Résumé : Cet article contribue à la compréhension des comportements économiques dans des contextes stratégiques inter-temporels. Par comparaison des équilibres markoviens parfaits et

Unit´e de recherche INRIA Rocquencourt, Domaine de Voluceau, Rocquencourt, BP 105, 78153 LE CHESNAY Cedex Unit´e de recherche INRIA Sophia-Antipolis, 2004 route des Lucioles, BP

To make the model tractable we also introduced a matching graph clustered using connected component clustering, that can be used to quickly organize very large image collections,