• Aucun résultat trouvé

ML estimation in the presence of a nuisance vector Using the Expectation Maximization (EM) algorithm Or the Bayesian Expectation Maximization (BEM) algorithm

N/A
N/A
Protected

Academic year: 2021

Partager "ML estimation in the presence of a nuisance vector Using the Expectation Maximization (EM) algorithm Or the Bayesian Expectation Maximization (BEM) algorithm"

Copied!
16
0
0

Texte intégral

(1)

HAL Id: hal-01410139

https://hal.archives-ouvertes.fr/hal-01410139

Submitted on 6 Dec 2016

HAL is a multi-disciplinary open access archive for the deposit and dissemination of sci- entific research documents, whether they are pub- lished or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.

ML estimation in the presence of a nuisance vector Using the Expectation Maximization (EM) algorithm Or

the Bayesian Expectation Maximization (BEM) algorithm

Jean-Pierre Cances

To cite this version:

Jean-Pierre Cances. ML estimation in the presence of a nuisance vector Using the Expectation Max- imization (EM) algorithm Or the Bayesian Expectation Maximization (BEM) algorithm. [Research Report] Xlim UMR CNRS 7252. 2009. �hal-01410139�

(2)

ML estimation in the presence of a nuisance vector Using the Expectation Maximization (EM) algorithm Or the Bayesian Expectation Maximization (BEM) algorithm

Cances Jean-Pierre (Octobre 2009), Xlim UMR 7252

The scope of this paper is to give a general framework to use two recent sophisticated algorithms for signal processing communication applications: EM and BEM algorithms.

Some contexts of application are then illustrated as application examples.

I. General Framework for EM algorithm :

We denote with r a random vector obtained by expanding the received modulated signal r(t) onto a suitable basis, and we indicate with b a deterministic vector of parameters to be estimated from the observation of the received vector r. Assume that r also depends on a random nuisance parameter vector a independent of b and with a priori probability density function (pdf) p( )a . The problem addressed here is to find the ML estimate ˆb of b, that is to say, the solution of:

 

ˆ arg max ln ( p )

b = b r b (1)

The likelihood function to be maximized with respect to the trial value bof b is obtained after elimination of the nuisance parameter vector a as follows:

( ) ( ) ( , )

p  p p d

a

r b a r a b a (2)

In order to solve (1), we take the derivative of ln (p r b) with respect to b and we equate it to zero, that is:

( ) ( , )

ln ( )

( ) ( , )

( , ) ( ). ( , )

0

( ) ( , )

p p d

p

p p d

p p p

d

p p

a

a

a

a r a b a

r b b

b a r a b a

r a b a r a b

b a

r b r a b

(3)

(3)

Using Bayes’ rule ( ). ( , )

( , )

( )

p p

p

p

a r a b

a r b

r b , we obtain the following equation:

ln ( ) ( , ) ln ( , ).

ln ( , ) , 0

p p p d

E p

 

 

a

a

r b a r b r a b a

b b

= r a b r b

b

(4)

In other words, the ML estimate ˆb of b is that value that nulls the conditional a posteriori expectation of the derivative with respect to b of the conditional log-likelihood function (LLF) ln (p r a b, ) .

Finding the solution of (4) is not trivial, since b appears in both factors of the integrand.

Thus, we try an iterative procedure that produces a sequence of values bˆ( )n hopefully converging to the desired solution. In particular, we use the previous sequence value bˆ(n1) to resolve the conditioning on the first factor of the integrand, and we find the current solution

ˆ( )n

b by solving the resulting simplified equation that follows:

( )

( -1)

( , ) ln ( , ) ˆn . 0

p n p d

b=b

a

a r b r a b a =

b (5)

If the sequence of estimates bˆ( )n yielded by (5) converges to a finite value, that value is a solution of ML equation (4) [1].

Observe now that the first factor of the integrand in (5) does not depend on bˆ( )n . Therefore, we can bring the derivative back out of the integral and obtain the equivalent equation:

( )

( ) ( -1)

ˆ : ( ,ˆ ) ln ( , ) ˆn 0

n n

p p d

 

 

a b=b

b a r b r a b a

b (6)

that is, the estimate bˆ( )n maximizes the conditional a posteriori expectation of the conditional LLF ln (p r a b, ) :

 

( ) ( -1)

ˆn arg max ( , ˆn )

b b b b (7-a)

( 1)

( -1)

( 1)

ˆ ˆ

( , ) ln ( , ) ,

( ,ˆ ) ln ( , )

n n

n

E p

p p d

 

a

a

b b r a b r b

a r b r a b a

(7-b)

Formulation (7-a)-(7-b) of our iterative solution can also be derived by means of the EM algorithm [2-4]. Consider r as the “incomplete” observation and z(rT,aT T) as the

“complete” observation. The EM algorithm states that the sequence bˆ( )n defined by:

- (i) expectation step (E-step)

(4)

 

( 1) ( 1)

ˆ ˆ

( , n ) ln ( ) , n

Q E p

b b a z b r b (8-a) - (ii) maximization step (M-step)

 

( ) ( 1)

ˆ n arg max Q( , ˆn )

b b b b (8-b)

converges to the ML estimate under mild conditions [2-3]. To make (8-a) (8-b) equivalent to (7-a) (7-b), we observe that, by using Bayes rule and considering that the distribution of a does not depend on the parameter vector to be estimated:

( ) ( , ) ( , ) ( )

( , ). ( )

p p p p

p p

z b r a b r a b a b r a b a

(9)

Therefore, substituting (9) in (8-a), we get :

( 1) ( -1) ( -1)

1 2

ˆ ˆ ˆ

( , n ) ( , n ) ln ( , ) ( , n ) ln ( )

Q p p d p p d

I I

 

a a

b b a r b r a b a a r b a a

(10)

The second term I2 in (10) does not depend on b, and as far the M-step is concerned, it can be dropped. Consequently, the estimation procedure given by (7-a) (7-b) and the EM algorithm defined by (8-a) (8-b), yield the same sequence of estimates. We explicitly observe that the solution of (1) can be found iteratively be only using a posteriori probabilities p(a r b,ˆ( -1)n ) and the LLF ln (p r a b, ) .

II. Application of EM to Synchronization for SISO-based receivers : EM-based synchronization

We show here how to apply the general framework of the previous section to the estimation of the synchronization parameters for a digital data-modulated bandpass signal. In this context, the nuisance parameter vector a contains the values of the N unknown (hence random) transmitted symbols, that is, aT a a0, ,1 ,aN1. Those symbols take values in an M-point constellation  (M-PSK, M-QAM,…). Thus, the vector a has a probability mass function (pmf) P(a), with  0, 1,,N1 and N. The vector b contains the synchronization parameters to be estimated, that is, bT A, , ,   where A, , ,   are the

channel gain, symbol timing, carrier frequency, and phase offsets, respectively. Here, the synchronization parameters are assumed as constant within the received code block. This has the advantage of simplifying notably the processing required by the estimation algorithm.

Furthermore, for the sake of simplicity, we will consider in the sequel an AWGN channel as well. Hence, the baseband received signal r(t) can be written as:

(5)

1 (2. . . ) 0

( ) .N k. ( . ). j t ( )

k

r t A a g t k T e    w t

(11)

where T is the symbol period, g(t) is a unit-energy (square-root raised-cosine) pulse, and w(t) is complex-valued AWGN with power spectral density 2.N0 (assumed to be known).

Neglecting irrelevant terms independent of a and b, the conditional LLF of (11) is

01 * .

2 01 2

ln ( , ) 2. .Re N k. ( , ).k j .N k

k k

p A a z v e A a

 

 

r a b (12)

where

.2. . .

.2. . .

.

( , ) ( ). ( . )

( ). ( )

j t

k

j t

t k T

z v r t e g t k T dt

r t e g t

 

 





 

 

(13)

is obtained by precompensating the received signal by the trial value , then applying the result to the matched filter g(-t), and finally sampling the matched filter output at the trial instant t k T. . Substituting (12) into (7-b) and dropping the terms which do not depend on b, we get:

1 *

( -1) ( -1) .

0

1 2

2 ( -1)

0

ˆ ˆ

( , ) 2. .Re ( , ) ( , ).

( ,ˆ )

n N n j

k k

k

N n

k k

A a p d z v e

A a p d

   

 

 

a

a

b b a r b a

a r b a

(14)

We now define

( -1) ( -1)

( -1)

ˆ ˆ

( , ) . ( , )

. ( ,ˆ )

m

n n

k k

n

m k m

a p d

P a



 

a

r b a r b a

r b

(15-a)

( 1) 2 ( -1)

2 ( -1)

ˆ ˆ

( , ) ( , )

. ( ,ˆ )

m

n n

k k

n

m k m

a p d

P a



 

a

r b a r b a

r b

(15-b)

( -1)

( k m ,ˆn )

P a r b denotes the marginal a posteriori probability (APP) of the kth channel symbol ak conditioned on the observation r and on the estimate bˆ( -1)n at the previous (n-1)th step, and m the M possible values taken in the constellation . Equation (14) can then be rearranged as:

(6)

1

( -1) * ( -1) .

0

2 1 ( 1)

0

ˆ ˆ

( , ) 2. .Re ( , ). ( , ).

( ,ˆ )

n N n j

k k

k

N n

k k

A z v e

A

 

 

b b r b

r b

(16)

We emphasize the similarity between (12) and (16): the latter is formally obtained from the former by simply replacing the terms ak and ak 2 by their respective a posteriori expected values k( ,r bˆ( -1)n ) and k( ,r bˆ(n1)). The new estimate bˆ( )n at the nth step is then determined by applying (7-a) and therefore by maximizing ( ,b b ˆ( -1)n ), given by (16) with respect to b. The corresponding result is:

( ) ( ) 1 * ( -1)

ˆ ˆ, 0

ˆ n ,ˆn arg max N k( ,ˆ n ). ( , )k

k

  z v

 

r b   (17-a)

ˆ( )n

 

01 * ( -1) ( ) ( )

( ,ˆ ). [ , ]

N n n n

k k

k

z v

r b (17-b)

1 * ( -1) ( ) ( )

( ) 0

1 ( 1)

0

( ,ˆ ). [ , ]

ˆ

( ,ˆ )

N n n n

k k

n k

N n

k k

z v A

r b r b

(17-c)

The obtained solution can be interpreted as an iterative synchronization procedure, which can be referred to as soft-decision-directed (SDD) synchronization. What we call here soft decisions are the a posteriori average values k( ,r bˆ( -1)n ) and k( ,r bˆ(n1)) of each channel symbol. They are a sort of “weighted average” over all the constellation points according to the respective symbol APPs. Note that thanks to (15-a) and (15-b), these a posteriori average values k( ,r bˆ( -1)n ) and k( ,r bˆ(n1))can be computed from marginals P a( k m r b,ˆ( -1)n ) only. In other words, due to the particular structure of the digital data-modulated signal, the implementation of the iterative ML estimation algorithm only requires the evaluation of the marginal a posteriori symbol probabilities P a( k m r b,ˆ( -1)n ).

We now concentrate on the evaluation of the marginal a posteriori symbol probabilities.

Whereas for uncoded transmission the usual assumption is that data symbols are independent and equally likely (yielding P(a)MN for all N), for a coded transmission with code rate , we only have a subset B N of all possible sequences corresponding toMN legitimate encoder output sequences. Therefore, taking into account that the APP of the symbol sequence a is given by:

( ). ( , )

( , )

( ). ( , )

B

P p

P

P p

a r a b

a r b

a r a b

(18)

and assuming that:

( ) ,

0,

M N B

P

B



a  

(19)

(7)

we get:

( , )

,

( , )

( , )

0,

B

p B

P p

B

 



r a = b r a b a r b

(20)

which relates the APP of the symbol sequence to the conditional likelihood function. Note that the result for uncoded transmission is obtained from (20) by taking B N. Finally, the marginal APP related to a symbol ak is obtained summing the symbol sequence APPs (20) over all symbols ai with i ≠ k. Evaluation of the APPs according to (20) yields a computational complexity that increases exponentially with the sequence length N, as all possible data sequences must be enumerated. However, in systems where the received signal can be modeled as a Markov process, the marginal symbol APPs can be efficiently obtained using the BCJR algorithm with a complexity that grows only linearly with the sequence length N.

Using a simple gate function for h(t) instead of the square-root raised-cosine filter, we can completely compute ( ,b b ˆ( -1)n ). We have:

.2. . .

( , ) ( ). j t. ( . ).

z vk r t e   h t k T dt



  (21)

with 0h t( )T, (21) is simplified in:

( 1).

.2. . . .

( , ) ( ).

k T

j t

k

k T

z v r t e dt

 

  (22)

Using r t( )A a e. .k j(2. . .  t ) and substituting into (22), we obtain:

( 2. .( ).( . )) . .( ). sin( .( ). )

( , ) . . . .

.( )

j k T j T

k k

z v A a e    e       T

  

  (23)

For the computation of k( ,r bˆ( -1)n ), we suppose that a simple BPSK is used and we get:

( -1) ( )

( ,ˆ ) tanh( )

2

n n k

k

r b L a (24)

with n denoting the nth corresponding turbo iteration. Then, combining (23) and (24) we eventually obtain:

1 *

( -1) .

0

1 .( ) (2. .( ).( . )) . .( ).

0

( , ) 2. .Re ( ,ˆ ) ( , ).

( ) sin( .( ). )

2. .Re tanh( ). . . . . .

2 .( )

N n j

L k k

k

N n k j j k T j T

k k

v A a p d z v e

L a T

A A a e e e

       

  

  

   

 

   

a

a r b a

(25)

(8)

Looking for a maximum value in (25) with tentative values   , constitutes a highly complicated task. One way of simplifying this problem would be to expand L( , )v  as a Fourier serie [5].

III. General Framework for BEM algorithm :

In order to introduce the key differences between EM and BEM, we recall the main properties of EM. Let   0, 1,...,L1Tdenote an L-dimensional deterministic vector to be estimated from an N-dimensional received vector R[R R0, 1,...,RN1]Tof noisy data (with N

≥ L). The ML estimation of is the solution of the problem [6]

arg max ( )

ML Lr

(26)

where Lr( ) log (f r ) is a log-likelihood function and f(x y) denotes the probability density function (pdf) of the random vector X conditioned on the event Y y. Solving the problem (26) in a direct fashion requires a closed form expression for Lr( ) but, even if this expression is available, the search for its maximum may entail an unacceptable computational burden. When this occurs, a feasible alternative can be offered by the EM algorithm [1-2].

The EM approach develops from the assumption that a complete data vector

0 1 1

[C C, ,...,CP ]T

C (with P ≥ N) is observed in place of the incomplete data set R. The vector C is characterized by a couple bof relevant properties: (1) it is not observed directly but, if available, would ease the estimation of ; (2) R can be obtained from C through a many-to-one mapping CR C( ). In practice, in communication problems, C is always chosen as a superset of the incomplete data [2], that is,

[ T, T T]

C R I (27)

where the so-called imputed data I are properly selected to simplify the ML estimation problem [1]. In particular, when consists of all transmitted channel symbols, I collects all the unwanted random parameters (fading, phase jitter, …) affecting the communication channel. These choices lead to hard detection algorithms often having an acceptable complexity and capable of incorporating the statistical properties of the channel parameters.

In the following, the complete data vector C will be always structured as in (27).

Given C, the auxiliary function :

 

 

( , ) ( ) ,

log ( ) ,

log ( ). ( , ).

i

EM c

Q E L

E f

f f d

 

I

I

S

R r

C R r

r,i i r i

 

(28)

(9)

is evaluated, where EX. denotes the statistical average with respect to a random vector X and Si is the space of I. The, this function is employed in the following two-step procedure generating successive approximations ( )k ,k 1, 2,...of ML (1):

(1) Expectation step-- QEM( , )  in (28) is evaluated for    EM( )k

(2) Maximization step-- given EM( )k , the next estimate EM(k1) is computed as:

( 1) ( )

arg max ( , ), 0,1,...

k k

EM QEM EM k

  (29)

An initial estimate EM(0) of  must be provided for the algorithm start-up. In digital communication problems, proper initialization of the EM algorithm is usually accomplished exploiting the information provided by known symbol pilots [1]. It can be proved that, under mild conditions, the sequence  EM( )k converges to the true ML estimate ML of (1), provided that the existence of local maxima does not prevent it. Avoiding this requires an accurate initial estimation EM(0) whose choice, for this reason, is of crucial importance [2].

The BEM algorithm : The unknown vector   0, 1,...,L1Tmentioned in the previous paragraph can also be modeled as a random quantity, when its joint pdf f() is available. In this case the MAP estimate MAP of , given the observed data vector r, can be evaluated as:

arg max ( )

MAP Mr

(30)

where Mr( ) log ( , )f r . Solving (30) remains a formidable task for the same reasons previously illustrated for the ML problem (1). In principle, however, an improved estimate of

can be evaluated via the MAP approach since statistical information about channel uncertainty are exploited.

Since there is a strong analogy between the ML problem (26) and the MAP one (30), it is not surprising that an expectation-maximization procedure, dubbed Bayesian EM (BEM) [7-8], for solving the latter, is available. The BEM algorithm evolves through the same iterative procedure, as the EM, but with a different auxiliary function [7], namely:

 

 

( , ) ( ) ,

log , ,

log ( , , ). ( ).

i

BEM C c

Q E M

E f

f f d

 

S

R r

C R r

r i i r, i

 

 

  (31)

A clear relationship can be established between the BEM And the EM algorithms. In fact, factoring the pdf f( , , )r i as:

( , , ) ( , ). ( )

f r i f r i f (32) and substituting (32) into (31) produces

(10)

( , ) ( , ) ( )

BEM EM

Q   Q   I (33)

where I( ) log ( )f . Equation (33) shows that the difference between QBEM( , )  (31) and ( , )

QEM   (28) is simply a bias term I( ) favoring the most likely values of . It is worth noting that, if a priori information about were unavailable and, consequently, a uniform pdf was selected for f( ) , the contribution from I( ) would turn into a constant in (33), that is, it could be neglected. Therefore, the BEM encompasses the EM as a special case and, since the former benefits by the statistical information about , it is expected to provide improved accuracy with respect to the latter. For the same reason, an increase in the speed of convergence and an improved robustness against the choice of the initial conditions could be offered by the BEM.

3.1 SISO Data detection in the presence of parametric uncertainty via the BEM Technique

In this section we show how the BEM technique can be employed to derive SISO algorithms for detecting digital signals transmitted over channels with parametric uncertainty and memory. A single-user transmission over a single input-single output channel is considered for simplicity, but the proposed approach can be extended to an arbitrary number of users and to MIMO system without any substantial conceptual problem.

Here, we assume that the kth component of the received data vector R can be expressed as:

( , )

k k k

R g D A N (34)

where D = [D D0, 1,...,DN1]T is a vector of independent channel symbols belonging to a constellation  s s0, ,...,1 sM1 of cardinality M and average energy Es,

0 1 1

[A A, ,...,AL]T

A is a vector of random channel parameters independent of D and with known statistical properties,  Nk is an AWGN sequence with variance N2, and gk(,., ) expresses the known functional dependence of the channel on both the transmitted symbols and its parametric uncertainty. In particular, we focus on conditional finite memory channels, that is, on random channels such that:

1 2

( , ) ( , , ,..., , )

k k k k k k Lc

g D A g D D D D A (35)

where Lc denotes the channel memory. The goal is to devise a MAP SISO detection algorithm given the observed data R = r and a statistically known parameter vector A. In data detection problems involving the EM technique, two different choices have been usually suggested for the imputed data I and the parameter vector :

(1) and =

(2) and =

I A D

I D A

(36)

It is extremely important to comment now on the meaning and the consequences of these choices. In the first case, both EM and BEM-based algorithms aim at producing hard estimates of the transmitted data. The only substantial difference between these two classes of

Références

Documents relatifs

We demonstrate how a global implicit function theorem can be used to verify such properties in the cases of finite mixtures of distributions in the exponential family, and

first case corresponds to a typical state and parameter estimation problem where the true stochas- tic parameter distribution is known, the second case corresponds to the estimation

In this paper we presented a novel direction of arrival estimation in the context of a noise mixture constituted of a Gaussian noise, which represents the thermal noise, and

Fast Incremental Expectation Maximization (FIEM) is an iterative algorithm, based on the Expectation Maximization (EM) algorithm, which was introduced to design EM for the large

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des

Keywords Deformable template · geometric variability · maximum likelihood estimation · missing variable · high dimension · stochastic EM algorithm · MCMC · Anisotropic MALA..

To this regard, we borrow from [11, 27] (see also [24] and the algorithm SARAH) a new technique called Stochastic Path-Integrated Differential Estimator (SPIDER) to generate the

In the expectation, or E-step, the missing data are estimated given the observed data and current estimate of the model parameters... achieved using the conditional