Rao-Blackwellised Particle Filtering For Dynamic Bayesian Networks
Rao-Blackwellised Particle Filtering For Dynamic Bayesian Networks
2000), fixed parameter estimation (Kong, Liu and Wong the alternative recursion
1994), HMMs (Doucet 1998, Doucet, Godsill and Andrieu
( p(Yt/Y1:t-l,ro:t)P(rt/rt-!)p(ro:t-11Yl:t-d
p ro:t I Yl:t) -
_
N
1
PN (ro,t,Xo:tl Yl:t) """' J( (i) (il ) (dro:tdXo:t)
=
N L.....t ro x
:t, O:t
=l i
w (ro:t) p(ro:tiYI:t)
t)
w (ro:t,Xo:t) = p=---:--((r....:o...:.:t-'x'-' o'-':t-:-1Y=-1'-: 7 q( ro:tl YI:t)
q ro:t. Xo:tl YI:t)
G.1ven .1.1. . d. sampIes { (r(0,ti),x0,t(i)) } q( ro:tl YI:t) J q( ro:t,Xo:t I YI:t) dxo:t
N d1" stn"buted accord-
ing to q( ro:t,Xo:tl YI:t), a Monte Carlo estimate of I (ft) Il:I._(ft) will require
Intuitively, to reach a given precision,
is given by a reduced number N of samples over I}v (ft) as we only
LiN=l ft (ro:t(i) ,xo:t(i)) w (ro:t(i),xo:t) (i) need to sample from a lower-dimensional distribution. This
is proven in the following propositions.
N
Li=l w (ro:n (i) Xo:t(i)) Proposition 1 The variances of the importance weights,
N the numerators and the denominators satisfy for any N
wa:t�t (ro:t
"" -(i)�
L.....t
(i),Xo:t(i)) (w (ro,t))::;
i=l Varq(r0,, ly1,t)
varq(ro,tiHt) )
(A� (ft) <
varq(r0,.,x0,, IYu) (W (ro,t, Xo,t))
varq(ro,.,xo,tiYt,tl (4 (ft) )
where the normalized importance weights wt� are equal to
(nJv (ft))
(i) (i)
varq(ro,tiYu) < varq(ro,.,xo,tiHtl (BJ; (ft))
WO:t(i) -- NW (rO:t'( (Xj)o:t)(j) IJv (ft) to
-
k=I
4 RAO-BLACKWELLISED PARTICLE
we can obtain recursive formulas to evaluate w (ro:t) =
. .
to the postenor p (r0(i),p Xo:t
(i} IYI:t) , at time
.
t. Th' ac- IS IS
[L::%1 w}1l ] - w} } . Hence we can perform importance
complished with the algorithm shown below, the details of
sampling online.
which will now be explained.
Choice of the Importance Distribution
Generic RBPF
There are infinitely many possible choices for q(ro:tI YI:t),
the only condition being that its supports must include that
(r�il) ,...., q(rtlr���-1 , Yl:t) most widely used distribution, since it is simple to compute,
but it can be inefficient, since it ignores the most recent
and set: evidence, Yt· Intuitively, many of our samples may end up
------'·
tions of the form (3), the variance of w (ro:t)
can only in
crease (stochastically) over time. The proof of this propo
sition is an extension of a Kong-Liu-Wong theorem (Kong
180 UNCERTAINTY IN ARTIFICIAL INTELLIGENCE PROCEEDINGS 2000
et a!. 1994, p. 285) to the case of an importance function of 4.2 CONVERGENCE RESULTS
the form (3).
Let B(IRn) be the space of bounded, Borel measurable
Proposition 4 The unconditional variance (i.e. with the
functions on IRn. We denote 11!11 �
sup If (x)l. The fol-
xEJRn
observations Y1:t being interpreted as random variables) lowing theorem is a straightforward consequence of Theo-
of the importance weights w (ro:t) increases over time. rem I in (Crisan and Doucet 2000) which is an extension
of previous results in (Crisan et a!. 1999).
In practice, the degeneracy caused by the variance increase Theorem 5 If the importance weights Wt are upper
can be observed by monitoring the importance weights. bounded and if one uses one of the selection schemes de
Typically, what we observe is that, after a few iterations, scribed previously, then, for all t 2: 0, there exists Ct inde-
one of the normalized importance weights tends to I, while pendent of N such that for any ft E B ( (JRnz )t+l )
the remaining weights tend to zero.
particles, var(N;). Recent theoretical results in (Crisan, any t) for the filtering distribution p ( Xt I Yl:t). We do not
Del Moral and Lyons 1999) indicate that the restriction pursue this here.
E( N;) N= w�i)
is unnecessary to obtain convergence re
sults (Doucet et a!. 1999). Examples of these selection 5 EXAMPLES
schemes include multinomial sampling (Doucet 1998, Gor
don et a!. 1993, Pitt and Shephard 1999), residual resam We now illustrate the theory by briefly describing two ap
pling (Kitagawa 1996, Liu and Chen 1998) and stratified plications we have worked on.
sampling (Kitagawa 1996). Their computational complex
ity is 0 (N).
5.1 ON-LINE REGRESSION AND MODEL
SELECTION WITH NEURAL NETWORKS
") -- ' --
UNCERTAINTY IN ARTIFICIAL INTELLIGENCE PROCEEDINGS 181
.
/··---.. ,.............
;------•{ � }------
/----....,
,./
1, r 'r _.L ,_-
J.
-
..
/ -, �- --
_ ,--, /·---,
: a0 � a1 ;'------'.---.{
' __ /
- .
' .. /
o' - -
hyper-parameters have been omitted for clarity. sensors are not perfect (they may accidentally flip bits),nor
are the motors (the robot may fail to move in the desired di
-
rection with some probability due e.g., to wheel slippage).
Consequently,it is easy for the robot to get lost. And when
�
.
.' the robot is lost,it does not know what part of the matrix to
��
£-t
·_. . ' . .
-\
.
.. .. . ..
update. So we are faced with a chicken-and-egg situation:
.
.
-2 the robot needs to know where it is to learn the map, but
230 240 250 260 270 280 290 300 310
needs to know the map to figure out where it is.
ts;;. � - �- � :.�.: 1
The problem of concurrent localization and map learn
ing for mobile robots has been widely studied. In (Mur
phy 2000), we adopt a Bayesian approach, in which we
0 50 100 150 200 250 300 350 400 450 500 maintain a belief state over both the location of the robot,
()
Lt E {1, . . . , N£ } ,and the color of each grid cell,Mt i E
�::� ; ; ; ' ; ; ; : 1
{1, . . . , Nc } i = 1, . . . , N£, where NL is the number
,
model. The middle and bottom plots show the true val
a function that flips its binary argument with some fixed
ues and estimates of the model order and noise variance
probability. In other words, the robot gets to see the color
respectively.
of the cell it is currently at, corrupted by noise: yt is a
noisy multiplexer with Lt acting as a "gate" node. Note
of basis functions online. For example, we generated some that this conditional independence is not obvious from the
data from a mixture of 2 RBFs for t = 1, ... , 500, and graph structure in Figure 3(a), which suggests that all the
then from a single RBF fort= 501, ... , 1000; the method nodes in each slice should be correlated by virtue of sharing
was able to track this change,as shown in Figure 2. Further a common observed child, as in a factorial HMM (Ghahra
experiments on real data sets are described in (Andrieu et mani and Jordan 1997). The extra independence informa
al. 1999a). tion is encoded in yt 's distribution, c.f., (Boutilier, Fried
man,Goldszmidt and Koller 1996).
5.2 ROBOT LOCALIZATION AND MAP The basic idea of the algorithm is to sample Ll:t with a PF,
BUILDING ()
and marginalize out the Mt i nodes exactly,which can be
done efficiently since they are conditionally independent
Consider a robot that can move on a discrete, two given Ll:t:
dimensional grid. Suppose the goal is to learn a map of
the environment, which, for simplicity, we can think of as P(Mt(l), ... , Mt(NL)iyl:t, Ll:t) = IJf�i P(Mt(i)iyl:t, Ll:t)
a matrix which stores the color of each grid cell, which
can be either black or white. The difficulty is that the color Some results on a simple one-dimensional grid world are
1 82 UNCERTAINTY IN ARTIFICIAL INTELLIGENCE PROCEEDINGS 2000
1 ... ll
admitting finite dimensional filters. That is, the filtering
\ / 11
il I
I '!"
distribution can be estimated in closed-form at any timet
using a fixed number of sufficient statistics. These include
l �
"'Wll
I
I
I �110 12 14 11
•
•
Dynamic models for counting observations (Smith
and Miller 1986).
RBPFs have been applied to many problems, mostly in Bernardo,J. M. and Smith, A. F. M. ( 1994). Bayesian The
the framework of conditionally linear Gaussian state-space ory, Wiley Series in Applied Probability and Statis
models and conditionally finite state-space HMMs. That is, tics.
they have been applied to models that, conditionally upon
Boutilier, C., Friedman, N., Goldszmidt, M. and Koller,
a set of variables (imputed by the PF algorithm), admit a
D. ( 1996). Context-specific independence in bayesian
closed-form filtering distribution (Kalman filter in the con
networks,Proc. Conf Uncertainty in AI.
tinuous case and HMM filter in the discrete case). One can
also make use of the special structure of the dynamic model Boyen, X. and Koller, D. ( 1998). Tractable inference
under study to perform the calculations efficiently using the for complex stochastic processes, Proc. Conf. Uncer
junction tree algorithm. For example, if one had evolv tainty in AI.
ing trees, one could sample the root nodes with the PF and
compute the leaves using the junction tree algorithm. This Casella, G. and Robert,C. P. ( 1996). Rao-Blackwellisation
would result in a substantial computational gain as one only of sampling schemes,Biometrika 83(1): 8 1-94.
has to sample the root nodes and apply the juction tree to
Cowell, R. G., Dawid, A. P., Lauritzen, S. L. and Spiegel
lower dimensional sub-networks.
halter, D. J. ( 1999). Probabilistic Networks and Ex
Although the previoulsy mentioned models are the most pert Systems, Springer-Verlag, New York.
UNCERTAINTY IN ARTIFICIAL INTELLIGENCE PROCEEDINGS 2000 183
Crisan, D. and Doucet, A. (2000). Convergence of gen Isard, M. and Blake, A. ( 1996). Contour tracking by
eralized particle filters, Technical Report CUEDIF stochastic propagation of conditional density, Euro
INFENG/TR 381, Cambridge University Engineering pean Conference on Computer Vision, Cambridge,
Department. UK, pp. 343-356.
Crisan, D., Del Moral, P. and Lyons, T. ( 1999). Dis Kanazawa,K.,Koller,D. and Russell,S. ( 1995). Stochastic
crete filtering using branching and interacting parti simulation algorithms for dynamic probabilistic net
cle systems, Markov Processes and Related Fields works, Proceedings of the Eleventh Conference on
5(3): 293-318. Uncertainty in Artificial Intelligence, Morgan Kauf
mann,pp. 346-35 1 .
de Freitas, J. F. G. ( 1999). Bayesian Methods for Neu
Kitagawa, G. ( 1996). Monte Carlo filter and smoother for
ral Networks, PhD thesis, Department of Engineer
non-Gaussian nonlinear state space models, Journal
ing, Cambridge University,Cambridge, UK.
ofComputational and Graphical Statistics 5: 1 -25.
Dean, T. and Kanazawa, K. ( 1989). A model for reason Kong, A., Liu, J. S. and Wong, W. H. (1994). Se
ing about persistence and causation,Artificial Intelli quential imputations and Bayesian missing data prob
gence 93(1 -2): 1 -27. lems,Journal of the American Statistical Association
89(425): 278-288.
Doucet, A. ( 1998). On sequential simulation-based meth
ods for Bayesian filtering,Technical ReportCUED/F Liu, J. S. and Chen, R. ( 1998). Sequential Monte Carlo
INFENG/TR 310, Department of Engineering, Cam methods for dynamic systems, Journal of the Ameri
bridge University. can Statistical Association 93: 1 032-1 044.
Doucet, A., de Freitas, J. F. G. and Gordon, N. J. MacEachern, S. N., Clyde, M. and Liu, J. S. ( 1999).
(2000). Sequential MonteCarlo Methods in Practice, Sequential importance sampling for nonparametric
Springer-Verlag. Bayes models: the next generation,Canadian Jour
nal of Statistics 27: 25 1 -267.
Doucet, A., Godsill, S. and Andrieu, C. (2000). On se
Murphy,K. P. (2000). Bayesian map learning in dynamic
quential Monte Carlo sampling methods for Bayesian
environments, in S. Solla, T. Leen and K.-R. Muller
filtering,Statistics andComputing 10(3): 197-208.
(eds), Advances in Neural Information Processing
Doucet, A., Gordon, N. J. and Krishnamurthy, V.
Systems 12, MIT Press, pp. 1 0 1 5- 1 02 1 .
( 1999). Particle filters for state estimation of jump Pearl, J. ( 1988). Probabilistic Reasoning in Intelligent Sys
Markov linear systems, Technical Report CUEDIF tems: Networks of Plausible Inference, Morgan Kauf
INFENG/TR 359, Cambridge University Engineering mann.
Department.
Pitt, M. K. and Shephard, N. ( 1999). Filtering via simula
Ghahramani, Z. and Jordan, M. ( 1997). Factorial Hidden tion: Auxiliary particle filters, Journal of the Ameri
Markov Models,Machine Learning 29: 245-273. can Statistical Association 94(446): 590-599.
Gilks, W. R. and Berzuini, C. ( 1998). Monte Carlo in Smith, R. L. and Miller, J. E. ( 1986). Predictive records,
ference for dynamic Bayesian models, Unpublished. Journal of the Royal Statistical Society B 36 : 79-88.
Medical Research Council, Cambridge, UK.
Uhlig, H. ( 1997). Bayesian vector-autoregressions with
stochastic volatility, Econometrica.
Gordon,N. 1., Salmond, D. J. and Smith, A. F. M. ( 1993).
Novel approach to nonlinear/non-Gaussian Bayesian Vidoni, P. ( 1999). Exponential family state space models
state estimation, lEE Proceedings-F 140(2): 1 07- based on a conjugate latent process, Journal of the
1 1 3. Royal Statistical Society B 61: 2 1 3-221 .
Green,P. J. ( 1995). Reversible jump Markov chain Monte West, M. ( 1993). Mixture models, Monte Carlo, Bayesian
Carlo computation and Bayesian model determina updating and dynamic models, Computing Science
tion,Biometrika 82: 7 1 1 -732. and Statistics 24: 325-333.
Handschin, J. E. and Mayne, D. Q. ( 1969). Monte Carlo West, M. and Harrison, J. ( 1996). Bayesian Forecasting
techniques to estimate the conditional expectation in and Dynamic Linear Models, Springer-Verlag.
multi-stage non-linear filtering,International Journal
ofControl 9(5): 547-559.