• Aucun résultat trouvé

Entropy Rate Revisited

Dans le document Entropy and Information Theory (Page 63-66)

Markov chainand note that givenZ,X does not depend onY. (Note that if Y →Z →X is a Markov chain, then so isX →Z→Y.) Thus the conditional mutual information is 0 if and only if the variables form a Markov chain with the conditioning variable in the middle. One might be tempted to infer from Lemma 2.3.3 that given finite valued measurementsf,g, andr

I(f(X);g(Y)|r(Z))(?)

≤I(X;Y|Z).

This does not follow, however, since it is not true that if Q is the partition corresponding to the three quantizers, thenD(Pf(X),g(Y),r(Z)||Pf(X)×g(Y)|r(Z)) is HPX,Y,Z||PX×Y|Z(f(X), g(Y), r(Z)) because of the way that PX×Y|Z is con-structed; e.g., the fact that X and Y are conditionally independent given Z implies thatf(X) andg(Y) are conditionally independent givenZ, but it does not imply thatf(X) andg(Y) are conditionally independent given r(Z). Al-ternatively, if M is PX×Z|Y, then it is not true that Pf(X)×g(Y)|r(Z) equals M(f gr)1. Note that if this inequality were true, choosing r(z) to be trivial (say 1 for all z) would result in I(X;Y|Z) I(X;Y|r(Z)) = I(X;Y). This cannot be true in general since, for example, choosing Z as (X, Y) would give I(X;Y|Z) = 0. Thus one must be careful when applying Lemma 2.3.3 if the measures and random variables are related as they are in the case of conditional mutual information.

We close this section with an easy corollary of the previous lemma and of the definition of conditional entropy. Results of this type are referred to aschain rulesfor information and entropy.

Corollary 2.5.1: Given finite alphabet random variables Y, X1, X2, · · ·, Xn,

H(X1, X2,· · ·, Xn) = Xn i=1

H(Xi|X1,· · ·, Xi1)

Hp||m(X1, X2,· · ·, Xn) = Xn i=1

Hp||m(Xi|X1,· · ·, Xi1)

I(Y; (X1, X2,· · ·, Xn)) = Xn

i=1

I(Y;Xi|X1,· · ·, Xi1).

2.6 Entropy Rate Revisited

The chain rule of Corollary 2.5.1 provides a means of computing entropy rates for stationary processes. We have that

1

nH(Xn) = 1 n

n1

X

i=0

H(Xi|Xi).

First suppose that the source is a stationarykth order Markov process, that is, for anym > k

Pr(Xn =xn|Xi=xi; i= 0,1,· · ·, n−1)

= Pr(Xn =xn|Xi=xi; i=n−k,· · ·, n−1).

For such a process we have for all n≥k that

H(Xn|Xn) =H(Xn|Xnkk) =H(Xk|Xk),

whereXim=Xi,· · ·, Xi+m−1. Thus taking the limit asn→ ∞of thenth order entropy, all but a finite number of terms in the sum are identical and hence the Ces`aro (or arithmetic) mean is given by the conditional expectation. We have therefore proved the following lemma.

Lemma 2.6.1: If{Xn}is a stationarykth order Markov source, then H¯(X) =H(Xk|Xk).

If we have a two-sided stationary process{Xn}, then all of the previous defi-nitions for entropies of vectors extend in an obvious fashion and a generalization of the Markov result follows if we use stationarity and the chain rule to write

1

nH(Xn) = 1 n

n1

X

i=0

H(X0|X1,· · ·, Xi).

Since conditional entropy is nonincreasing with more conditioning variables ((2.22) or Lemma 2.5.2),H(X0|X1,· · ·, Xi) has a limit. Again using the fact that a Ces`aro mean of terms all converging to a common limit also converges to the same limit we have the following result.

Lemma 2.6.2: If{Xn}is a two-sided stationary source, then H¯(X) = lim

n→∞H(X0|X1,· · ·, Xn).

It is tempting to identify the above limit as the conditional entropy given the infinite past,H(X0|X−1,· · ·). Since the conditioning variable is a sequence and does not have a finite alphabet, such a conditional entropy is not included in any of the definitions yet introduced. We shall later demonstrate that this interpretation is indeed valid when the notion of conditional entropy has been suitably generalized.

The natural generalization of Lemma 2.6.2 to relative entropy rates unfor-tunately does not work because conditional relative entropies are not in general monotonic with increased conditioning and hence the chain rule does not imme-diately yield a limiting argument analogous to that for entropy. The argument does work if the reference measure is akth order Markov, as considered in the following lemma.

2.6. ENTROPY RATE REVISITED 43 Lemma 2.6.3: If{Xn}is a source described by process distributionspand mand ifpis stationary andmiskth order Markov with stationary transitions, then forn≥k Hp||m(X0|X−1,· · ·, Xn) is nondecreasing innand

H¯p||m(X) = lim

n→∞Hp||m(X0|X−1,· · ·, Xn)

=−H¯p(X)−Ep[lnm(Xk|Xk)].

Proof: Forn≥kwe have that

Hp||m(X0|X−1,· · ·, Xn)

=−Hp(X0|X1,· · ·, Xn)X

xk+1

pXk+1(xk+1) lnmXk|Xk(xk|xk).

Since the conditional entropy is nonincreasing with nand the remaining term does not depend onn, the combination is nondecreasing withn. The remainder of the proof then parallels the entropy rate result. 2

It is important to note that the relative entropy analogs to entropy properties often requirekth order Markov assumptions on the reference measure (but not on the original measure).

Markov Approximations

Recall that the relative entropy rate ¯Hp||m(X) can be thought of as a distance between the process with distributionpand that with distributionmand that the rate is given by a limit if the reference measuremis Markov. A particular Markov measure relevant to p is the distribution p(k) which is the kth order Markov approximation to p in the sense that it is a kth order Markov source and it has the samekth order transition probabilities asp. To be more precise, the process distributionp(k) is specified by its finite dimensional distributions

p(k)Xk(xk) =pXk(xk)

p(k)Xn(xn) =pXk(xk)

nY1 l=k

pXl|Xk

l−k(xl|xklk); n=k, k+ 1,· · · so that

p(k)Xk|Xk =pXk|Xk.

It is natural to ask how good this approximation is, especially in the limit, that is, to study the behavior of the relative entropy rate ¯Hp||p(k)(X) ask→ ∞.

Theorem 2.6.2: Given a stationary processp, letp(k)denote thekth order Markov approximations top. Then

klim→∞H¯p||p(k)(X) = inf

k H¯p||p(k)(X) = 0.

Thus the Markov approximations are asymptotically accurate in the sense that the relative entropy rate between the source and approximation can be made arbitrarily small (zero if the original source itself happens to be Markov).

Proof: As in the proof of Lemma 2.6.3 we can write forn≥kthat Hp||p(k)(X0|X−1,· · ·, Xn)

=−Hp(X0|X−1,· · ·, Xn)X

xk+1

pXk+1(xk+1) lnpXk|Xk(xk|xk)

=Hp(X0|X−1,· · ·, Xk)−Hp(X0|X−1,· · ·, Xn).

Note that this implies thatp(k)Xn>> pXn for allnsince the entropies are finite.

This automatic domination of the finite dimensional distributions of a measure by those of its Markov approximation will not hold in the general case to be encountered later, it is specific to the finite alphabet case. Taking the limit as n→ ∞gives

H¯p||p(k)(X) = lim

n→∞Hp||p(k)(X0|X−1,· · ·, Xn)

=Hp(X0|X−1,· · ·, Xk)−H¯p(X).

The corollary then follows immediately from Lemma 2.6.2. 2

Markov approximations will play a fundamental role when considering rela-tive entropies for general (nonfinite alphabet) processes. The basic result above will generalize to that case, but the proof will be much more involved.

Dans le document Entropy and Information Theory (Page 63-66)