• Aucun résultat trouvé

Entropy Rate

Dans le document Entropy and Information Theory (Page 53-57)

MΦ(ω) = M(ω)eΦ(ω)

using the divergence inequality. Since this is true for any Φ, it is also true for the supremum over Φ and the theorem is proved. 2

2.4 Entropy Rate

Again let{Xn;n= 0,1,· · ·}denote a finite alphabet random process and apply Lemma 2.3.2 to vectors and obtain

H(X0, X1,· · ·, Xn1)

H(X0, X1,· · ·, Xm1) +H(Xm, Xm+1,· · ·, Xn1); 0< m < n. (2.18) Define as usual the random vectors Xkn = (Xk, Xk+1,· · ·, Xk+n−1), that is, Xkn is a vector of dimension n consisting of the samples of X from k to k+n−1. If the underlying measure is stationary, then the distributions of the random vectors Xkn do not depend on k. Hence if we define the sequence h(n) =H(Xn) =H(X0,· · ·, Xn−1), then the above equation becomes

h(k+n)≤h(k) +h(n); allk, n >0.

Thus h(n) is a subadditive sequence as treated in Section 7.5 of [50]. A basic property of subadditive sequences is that the limith(n)/nasn→ ∞exists and equals the infimum of h(n)/n over n. (See, e.g., Lemma 7.5.1 of [50].) This immediately yields the following result.

Lemma 2.4.1: If the distribution m of a finite alphabet random process {Xn}is stationary, then

Thus the limit exists and equals the infimum.

The next two properties of entropy rate are primarily of interest because they imply a third property, the ergodic decomposition of entropy rate, which

will be described in Theorem 2.4.1. They are also of some independent interest.

The first result is a continuity result for entropy rate when considered as a func-tion or funcfunc-tional on the underlying process distribufunc-tion. The second property demonstrates that entropy rate is actually an affine functional (both convexS and convexT

) of the underlying distribution, even though finite order entropy was only convexT

and not affine.

We apply the distributional distance described in Section 1.8 to the standard sequence measurable space (Ω,B) = (AZ+,BAZ+) with aσ-field generated by the countable fieldF={Fn; n= 1,2,· · ·}generated by all thin rectangles.

Corollary 2.4.1: The entropy rate ¯Hm(X) of a discrete alphabet random process considered as a functional of stationary measures is upper semicontinu-ous; that is, if probability measuresmand mn, n= 1,2,· · · have the property thatd(m, mn)0 asn→ ∞, then

H¯m(X)lim sup

n→∞

H¯mn(X).

Proof: For each fixedn Hm(Xn) = X

anAn

m(Xn=an) lnm(Xn=an)

is a continuous function ofmsince for the distance to go to zero, the probabilities of all thin rectangles must go to zero and the entropy is the sum of continuous real-valued functions of the probabilities of thin rectangles. Thus we have from Lemma 2.4.1 that ifd(mk, m)→0, then

The next lemma uses Lemma 2.3.4 to show that entropy rates are affine functions of the underlying probability measures.

Lemma 2.4.2: Letmandpdenote two distributions for a discrete alphabet random process{Xn}. Then for anyλ∈(0,1),

2.4. ENTROPY RATE 33 Ifmandpare stationary then

H¯λm+(1λ)p(X) =λH¯m(X) + (1−λ) ¯Hp(X) (2.21) and hence the entropy rate of a stationary discrete alphabet random process is an affine function of the process distribution. 2

Comment: Eq. (2.19) is simply Lemma 2.3.4 applied to the random vectors Xn stated in terms of the process distributions. Eq. (2.20) states that if we look at the limit of the normalized log of a mixture of a pair of measures when one of the measures governs the process, then the limit of the expectation does not depend on the other measure at all and is simply the entropy rate of the driving source. Thus in a sense the sequences produced by a measure are able to select the true measure from a mixture.

Proof: Eq. (2.19) is just Lemma 2.3.4. Dividing byn and taking the limit asn→ ∞proves that entropy rate is affine. Similarly, take the limit supremum in expressions (2.12) and (2.13) and the lemma is proved. 2

We are now prepared to prove one of the fundamental properties of entropy rate, the fact that it has an ergodic decomposition formula similar to property (c) of Theorem 1.8.2 when it is considered as a functional on the underlying distribution. In other words, the entropy rate of a stationary source is given by an integral of the entropy rates of the stationary ergodic components. This is a far more complicated result than property (c) of the ordinary ergodic decompo-sition because the entropy rate depends on the distribution; it is not a simple function of the underlying sequence. The result is due to Jacobs [68].

Theorem 2.4.1: The Ergodic Decomposition of Entropy Rate Let (AZ+,B(A)Z+, m, T) be a stationary dynamical system corresponding to a stationary finite alphabet

source {Xn}. Let {px} denote the ergodic decomposition of m. If ¯Hpx(X) is m-integrable, then

H¯m(X) = Z

dm(x) ¯Hpx(X).

Proof: The theorem follows immediately from Corollary 2.4.1 and Lemma 2.4.2 and the ergodic decomposition of semi-continuous affine funtionals as in Theorem 8.9.1 of [50]. 2

Relative Entropy Rate

The properties of relative entropy rate are more difficult to demonstrate. In particular, the obvious analog to (2.18) does not hold for relative entropy rate without the requirement that the reference measure by memoryless, and hence one cannot immediately infer that the relative entropy rate is given by a limit for stationary sources. The following lemma provides a condition under which the relative entropy rate is given by a limit. The condition, that the dominating measure be a kth order (ork-step) Markov source will occur repeatedly when dealing with relative entropy rates. A source is kth order Markov or k-step Markov (or simply Markov if k is clear from context) if for any n and any

N ≥k

P(Xn =xn|Xn−1=xn−1,· · ·, XnN =xnN)

=P(Xn =xn|Xn1=xn1,· · ·, Xnk =xnk);

that is, conditional probabilities given the infinite past depend only on the most recent k symbols. A 0-step Markov source is a memoryless source. A Markov source is said to havestationary transitionsif the above conditional probabilities do not depend onn, that is, if for anyn

P(Xn =xn|Xn1=xn1,· · ·, XnN =xnN)

=P(Xk=xn|Xk−1=xn−1,· · ·, X0=xnk).

Lemma 2.4.3 If pis a stationary process and m is a k-step Markov process with stationary transitions, then

H¯p||m(X) = lim

n→∞

1

nHp||m(Xn) =−H¯p(X)−Ep[lnm(Xk|Xk)], whereEp[lnm(Xk|Xk)] is an abbreviation for

Ep[lnm(Xk|Xk)] = X

xk+1

pXk+1(xk+1) lnmXk|Xk(xk|xk).

Proof: If for anynit is not true thatmXn >> pXn, thenHp||m(Xn) = for that and all largernand both sides of the formula are infinite, hence we assume that all of the finite dimensional distributions satisfy the absolute continuity relation. Sincemis Markov,

mXn(xn) =

nY−1 l=k

mXl|Xl(xl|xl)mXk(xk).

Thus 1

nHp||m(Xn) =1

nHp(Xn) 1 n

X

xn

pXn(xn) lnmXn(xn)

=1

nHp(Xn) 1 n

X

xk

pXk(xk) lnmXk(xk)

−n−k n

X

xk+1

pXk+1(xk+1) lnmXk|Xk(xk|xk).

Taking limits then yields

H¯p||m(X) =−H¯pX

xk+1

pXk+1(xk+1) lnmXk|Xk(xk|xk),

where the sum is well defined because if mXk|Xk(xk|xk) = 0, then so must pXk+1(xk+1) = 0 from absolute continuity. 2

2.5. CONDITIONAL ENTROPY AND INFORMATION 35

Dans le document Entropy and Information Theory (Page 53-57)