maximum a posteriori sequence estimation using monte carlo particle filters

18
Maximum Maximum a a posteriori posteriori sequence estimation sequence estimation using Monte Carlo using Monte Carlo particle filters particle filters S. J. Godsill, A. Doucet, and M. West Annals of the Institute of Statistical Mathem atics Vol. 52, No. 1, 2001. 조 조 조

Upload: hashim-powers

Post on 31-Dec-2015

21 views

Category:

Documents


1 download

DESCRIPTION

Maximum a posteriori sequence estimation using Monte Carlo particle filters. S. J. Godsill, A. Doucet, and M. West Annals of the Institute of Statistical Mathematics Vol. 52, No. 1, 2001. 조 동 연. Abstract. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Maximum  a posteriori  sequence estimation using Monte Carlo particle filters

Maximum Maximum aa posterioriposteriori sequence estimation using sequence estimation using Monte Carlo particle filtersMonte Carlo particle filters

S. J. Godsill, A. Doucet, and M. West

Annals of the Institute of Statistical Mathematics Vol. 52, No. 1, 2001.

조 동 연

Page 2: Maximum  a posteriori  sequence estimation using Monte Carlo particle filters

AbstractAbstract

Performing maximum a posteriori (MAP) sequence estimation in non-linear non-Gaussian dynamic models A particle cloud representation of the filtering distribution

which evolve through time using importance sampling and resampling ideas

MAP sequence estimation is then performed using a classical dynamic programming technique applied to the discretised version of the state space.

Page 3: Maximum  a posteriori  sequence estimation using Monte Carlo particle filters

IntroductionIntroduction

Standard Markovian state-space model

xt RRnx: unobserved states of the systems

yt RRny: observations made over some time interval

f(.|.) and g(.|.): pre-specified densities which may be non-Gaussian and involve non-linearity

f(x1| x0) f(x0)

x1:t, y1:t: collections of observations and states

densityn Observatio )|(~

densityevolution State )|(~ 1

ttt

ttt

xygy

xxfx

Page 4: Maximum  a posteriori  sequence estimation using Monte Carlo particle filters

Joint distribution of states and observations Markov assumptions

Recursion for this joint distribution

Computing this can only be performed in closed form for linear Gaussian models using the Kalman filer-smoother and for finite state space hidden Markov models.

Approximate numerical techniques

t

iiiiitt xygxxfyxp

11:1:1 )|()|()|(

)|(

)|()|()|()|(

1

111:1:11:11:1

tt

tttttttt yyp

xxfxygyxpyxp

Page 5: Maximum  a posteriori  sequence estimation using Monte Carlo particle filters

Monte Carlo particle filters Randomized adaptive grid approximation where the par

ticles evolve randomly in time according to a simulation-based rule

x0(dx): the Dirac delta function located at x0

wt(i): the weight attached to particle x(i)

1:t, wt(i) 0 and wt

(i) =1

Particles at time t can be updated efficiently to particles at time t+1 using sequential importance sampling and resampling.

Severe depletion of samples over time There are only a few distinct paths.

N

itx

ittt dxwyxp i

t1

:1)(

:1:1 )()|( )(:1

Page 6: Maximum  a posteriori  sequence estimation using Monte Carlo particle filters

MAP estimation Estimation of the MAP sequence

Marginal fixed-lag MAP sequence

For many applications, it is important to capture the sequence-specific interactions of the states over time in order to make successful inferences.

)|(maxarg)( :1:1:1:1

ttx

MAPt yxptx

t

)|(maxarg)( :1:1:1:1

ttLtx

MMAPtLt yxptx

tLt

Page 7: Maximum  a posteriori  sequence estimation using Monte Carlo particle filters

Maximum a Posteriori Maximum a Posteriori sequence estimationsequence estimation Standard methods

Simple sequential optimization method Sampling (sequentially in time) some paths according to a dist

ribution q(x1:t)

The choice of q(x1:t) will have a huge influence on the performance of the algorithm and the construction of an “optimal” distribution q(x1:t) is clearly very difficult.

A reasonable choice for q(x1:t) is the posterior distribution p(x1:t

| y1:t) or any distribution that has the same global maxima.

)|(maxarg)(ˆ :1:1},...,1;{

:1)(

:1:1

ttNixx

MAPt yxptx

itt

Page 8: Maximum  a posteriori  sequence estimation using Monte Carlo particle filters

A clear advantage of this method It is very easy to implement and has computational complexity

and storage requirements of order O(NT).

A severe drawback Because of the degeneracy phenomenon, the performance of

this estimate will get worse as time t increase. A huge number of trajectories is required for reasonable

performance, especially for large datasets.

Page 9: Maximum  a posteriori  sequence estimation using Monte Carlo particle filters

Optimization via dynamic programming Maximization of p(x1:t|y1:t)

The function to maximize is additive.

)|(maxarg)(ˆ :1:1},...,1;{

:1)(

1:1

ttNixx

MAPt yxptx

ik

tkt

t

kkkkk

x

MAPt xxfxygtx

t 11:1 )|(log)|(logmaxarg)(

:1

t

kkkkk

Nixx

MAPt xxfxygtx

ik

tkt 1

1},...,1;{

:1 )|(log)|(logmaxarg)(ˆ)(

1:1

Page 10: Maximum  a posteriori  sequence estimation using Monte Carlo particle filters

Viterbi algorithm

TMAPt

MAPMAPMAPt txtxtxtx ))(ˆ),...,(ˆ),(ˆ()(ˆ 21:1

Page 11: Maximum  a posteriori  sequence estimation using Monte Carlo particle filters

Maximization of p(xx-L+1:t|y1:t) The algorithm proceeds exactly as before, but starting a time t-L+1

and replacing the initial state distribution with p(xx-L+1:t|y1:t-L).

Computational complexity: O(N2(L+1)) Memory requirements: O(N(L+1))

Page 12: Maximum  a posteriori  sequence estimation using Monte Carlo particle filters

ExamplesExamples

A non-linear time series

tt

t

tt

ttt

wx

y

vtx

xxx

20

)2.1cos(81

252

1

2

21

11

)1,0(~ ),10,0(~ ),5,0(~1 NwNvNx k

Page 13: Maximum  a posteriori  sequence estimation using Monte Carlo particle filters

Simulated state sequence

Observations

Page 14: Maximum  a posteriori  sequence estimation using Monte Carlo particle filters

Filtering distribution p(xt|y1:t) at time t=14

Page 15: Maximum  a posteriori  sequence estimation using Monte Carlo particle filters

Evolution of the filtering distribution p(xt|y1:t) over time t

Page 16: Maximum  a posteriori  sequence estimation using Monte Carlo particle filters

Simulated sequence (solid)

MMSE estimate (dotted)

MAP sequence estimate (dashed)

Page 17: Maximum  a posteriori  sequence estimation using Monte Carlo particle filters

Comparisons Mean log-posterior values of the MAP estimate over 10 data

realization

Sample mean log-posterior values and standard deviation over 25 simulations with the same data

Page 18: Maximum  a posteriori  sequence estimation using Monte Carlo particle filters

Viterbi algorithm outperforms the standard method and that the robustness in terms of sample variability improves as the number of particles increases.

Because of the degeneracy phenomenon inherent in the standard method, this improvement over the standard methods will get larger and larger as t increases.