• Aucun résultat trouvé

The Theory behind PageRank

N/A
N/A
Protected

Academic year: 2022

Partager "The Theory behind PageRank"

Copied!
22
0
0

Texte intégral

(1)

The Theory behind PageRank

Mauro Sozio

Telecom ParisTech

May 21, 2014

(2)

Events and Probability

Consider a stochastic process (e.g. throw a dice, pick a card from a deck) Each possible outcome is a simple event.

The sample space Ω is the set of all possible simple events.

An event is a set of simple events (a subset of the sample space).

With each simple event E we associate a real number 0≤P(E)≤1, which is the probability that event E happens.

(3)

Probability Space

Definition

A probability spacehas three components:

Asample space Ω, which is the set of all possible outcomes of the random process modeled by the probability space;

A family of sets F representing the allowable events, where each set in F is a subset of the sample space in Ω;

aprobability function P :F →R, satisfying the definition below (next slide).

(4)

Probability Function

Definition

A probability function is any functionP :F →Rthat satisfies the following conditions:

for any eventE, 0≤P(E)≤1;

P(Ω) = 1;

for any finite or countably infinite sequence of pairwise mutually disjoint eventsE1,E2,E3, . . .

P(∪i≥1Ei) =X

P(Ei). (1)

(5)

The Union Bound

Theorem

P(∪ni=1Ei

n

X

i=1

P(Ei)). (2)

Example: roll a dice:

let E1 = “result is odd”

let E2 = “result is ≤2”

(6)

Independent Events

Definition

Two events E1 and E2 are independentif and only if

P(E1∩E2) =P(E1)·P(E2) (3)

(7)

Conditional Probability: Example

What is the probability that a random student at Telecom ParisTech was born in Paris?

E1 = the event “born in Paris”.

E2 = the event “student at Telecom ParisTech”.

The conditional probability that a a student at Telecom ParisTech was born in Paris is written:

P(E1|E2).

(8)

Conditional Probability: Definition

Definition

The conditional probabilitythat event E1 occurs given that eventE2

occurs is

P(E1|E2) = P(E1∩E2)

P(E2) (4)

The conditional probability is only well-defined if P(E2)>0.

By conditioning on E2 we restrict the sample space to the set E2.

(9)

Random Variable

Definition

A random variable X on a sample space Ω is a function on Ω; that is, X : Ω→R.

A discrete random variable is a random variable that takes on only a finite number of values.

(10)

Examples

In practice, a random variable is some random quantity that we are interested in:

I roll a die,X = result. E.g. X = 6.

I pick a card,X = 1 if card is an Ace, 0 otherwise.

I roll a dice two times. X1= result of the first experiment, X2 = result of the second experiment. What is P(X1+X2 = 2)?

(11)

Stochastic Processes

Definition

A stochastic process in discrete time n∈N is a sequence of random variables X0,X1,X2. . . denoted byX={Xn}.

We refer to the valueXn as the state of the process at timen, with X0

denoting the initial state.

The set of possible values that each random variable can take is denoted byS. Here, we shall assume thatS is finite and S ⊆N.

(12)

Markov Chains

Definition

A stochastic process{Xn} is called aMarkov chain if for anyn≥0 and any value j0,j1, . . . ,i,j ∈S,

P(Xn+1=i|Xn=j,Xn−1=jn−1, . . . ,X0 =j0) =P(Xn+1 =i|Xn=j), which we denote by Pij.

This can be stated as the future is independent of the past given the present state. In other words, the probability of moving to the next state

(13)

One-step Transition Matrix

Pij denotes the probability that the chain, whenever in statej, moves next into state i.

The square matrix P= (Pij), i,j ∈S, is called the one-step transition matrix. Note that for each j ∈S we have:

X

i∈S

Pij = 1. (5)

(14)

n-step Transition Matrix

The n-step transition matrixP(n),n≥1, where

Pijn=P(Xn=i|X0=j) =P(Xm+n=i|Xm =j), ∀m (6) denotes the probability that n steps later the Markov chain will be in state i given that at stepm is in statej.

Theorem

P(n) =Pn =P×P× · · · ×P,n≥1.

(15)

Definition

A Markov chain is called irreduciblea iff for anyi,j ∈S, there isn≥1 such that:

Pijn>0. (7)

adefinition is different whenS is not finite.

In other words, the chain is able to move from any state i to any statej (in one or more steps). As a result, if a Markov chain is irreducible then there must be n such that Piin>0.

(16)

Aperiodicity

A statei has period k if any return toi occurs at step k·l, for some l >0. Formally,

k =gcd{n :P(Xn=i|x0=i)>0,} (8) where gcd denotes the greatest common divisor. If k = 1 then statei is said to be aperiodic.

Definition

A Markov chain is called aperiodicif every state is aperiodic.

(17)

Stationary Distribution

Definition

A probability distributionπ over the states of the Markov chain (P

j∈Sπj = 1) is called astationary distributionif

π =πP. (9)

(18)

Main Theorem

Theorem

If a Markov chain is irreducible and aperiodica, then a stationary

distribution π exists and is unique. Moreover, the Markov chain converges to its stationary distribution, that is,

πj = lim

n→∞P(Xn=j) =P(Xn=j|X0 =i), ∀i,j ∈S. (10)

ain this case the Markov chain is calledergodic

Note that Equation (10) holds regardless the initial state i of the Markov

(19)

Markov Chains

Markov chains and the Random Surfer

Consider the Markov chain obtained from the web graph, no rand. jumps.

Is it irreducible?

Spider traps? Other questions:

What if we add random jumps?

Can we compute the probability distribution that an ergodic Markov chain converges to? How?

(20)

Markov Chains

Markov chains and the Random Surfer

Consider the Markov chain obtained from the web graph, no rand. jumps.

Is it irreducible?

Is it aperiodic?

Other questions:

What if we add random jumps?

Can we compute the probability distribution that an ergodic Markov chain converges to? How?

(21)

Markov Chains

Markov chains and the Random Surfer

Consider the Markov chain obtained from the web graph, no rand. jumps.

Is it irreducible?

Is it aperiodic?

Spider traps?

Other questions:

What if we add random jumps?

chain converges to? How?

(22)

Markov chains and the Random Surfer

Consider the Markov chain obtained from the web graph, no rand. jumps.

Is it irreducible?

Is it aperiodic?

Spider traps?

Other questions:

What if we add random jumps?

Can we compute the probability distribution that an ergodic Markov chain converges to? How?

Références

Documents relatifs

Any collection of two-cycles, triangles, and at most one path with the same orientation as the the triangles; with in total T − 1 edges, is a decomposition of the state graph of

Economic diplomacy can roughly and briefly be understood as 'the management of relations between states that affect international economic issues intending to

Remark: Given a HMW

We can mention, for instance, the Back and Forth Nudging proposed by Auroux and Blum [1], the Time Reversal Focusing by Phung and Zhang [10], the algorithm proposed by Ito, Ramdani

This paper introduces a new model for the SoC estimation using a Switching Markov State-Space Model (SMSSM): the battery behavior is described by a set of potential linear state-

And if one’s view is that something plays the role of S’s motivating reason just in case (a) S takes it to speak in favor of Φing, (b) is motivated by it to Φ, (c) is motivated by

Entre 1998 et 2004, dans les entreprises de 10 salariés ou plus du secteur marchand non agricole, la durée annuelle col- lective du travail a diminué de 1 742 à 1 616 heures en

Sen, Spectral norm of random large dimensional noncentral Toeplitz and Hankel matrices, Electron... Yin, Limit of the smallest eigenvalue of a large-dimensional sample