Continuous Variables Write message update equation as an expectation: Proposal distribution W t (x t ) for each node Samples define a random discretization.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Bayesian Belief Propagation
Linear Time Methods for Propagating Beliefs Min Convolution, Distance Transforms and Box Sums Daniel Huttenlocher Computer Science Department December,
Exact Inference. Inference Basic task for inference: – Compute a posterior distribution for some query variables given some observed evidence – Sum out.
Exact Inference in Bayes Nets
ICCV 2007 tutorial Part III Message-passing algorithms for energy minimization Vladimir Kolmogorov University College London.
Dynamic Bayesian Networks (DBNs)
Loopy Belief Propagation a summary. What is inference? Given: –Observabled variables Y –Hidden variables X –Some model of P(X,Y) We want to make some.
3D Human Body Pose Estimation from Monocular Video Moin Nabi Computer Vision Group Institute for Research in Fundamental Sciences (IPM)
Belief Propagation by Jakob Metzler. Outline Motivation Pearl’s BP Algorithm Turbo Codes Generalized Belief Propagation Free Energies.
Markov Networks.
Introduction to Sampling based inference and MCMC Ata Kaban School of Computer Science The University of Birmingham.
CS774. Markov Random Field : Theory and Application Lecture 04 Kyomin Jung KAIST Sep
Distributed Message Passing for Large Scale Graphical Models Alexander Schwing Tamir Hazan Marc Pollefeys Raquel Urtasun CVPR2011.
Belief Propagation in a Continuous World Andrew Frank 11/02/2009 Joint work with Alex Ihler and Padhraic Smyth TexPoint fonts used in EMF. Read the TexPoint.
1 Graphical Models in Data Assimilation Problems Alexander Ihler UC Irvine Collaborators: Sergey Kirshner Andrew Robertson Padhraic Smyth.
CS 188: Artificial Intelligence Fall 2009 Lecture 20: Particle Filtering 11/5/2009 Dan Klein – UC Berkeley TexPoint fonts used in EMF. Read the TexPoint.
Graphical Models for Mobile Robot Localization Shuang Wu.
Global Approximate Inference Eran Segal Weizmann Institute.
Machine Learning CUNY Graduate Center Lecture 7b: Sampling.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Probabilistic Robotics
Particle filters (continued…). Recall Particle filters –Track state sequence x i given the measurements ( y 0, y 1, …., y i ) –Non-linear dynamics –Non-linear.
Learning Low-Level Vision William T. Freeman Egon C. Pasztor Owen T. Carmichael.
Understanding Belief Propagation and its Applications Dan Yuan June 2004.
Today Introduction to MCMC Particle filters and MCMC
Comparative survey on non linear filtering methods : the quantization and the particle filtering approaches Afef SELLAMI Chang Young Kim.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
The Role of Specialization in LDPC Codes Jeremy Thorpe Pizza Meeting Talk 2/12/03.
Image Analysis and Markov Random Fields (MRFs) Quanren Xiong.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Computer vision: models, learning and inference
Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei.
6. Experimental Analysis Visible Boltzmann machine with higher-order potentials: Conditional random field (CRF): Exponential random graph model (ERGM):
Model Inference and Averaging
Machine Learning Lecture 23: Statistical Estimation with Sampling Iain Murray’s MLSS lecture on videolectures.net:
Recap: Reasoning Over Time  Stationary Markov models  Hidden Markov models X2X2 X1X1 X3X3 X4X4 rainsun X5X5 X2X2 E1E1 X1X1 X3X3 X4X4 E2E2 E3E3.
Probabilistic Robotics Bayes Filter Implementations.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Mixture Models, Monte Carlo, Bayesian Updating and Dynamic Models Mike West Computing Science and Statistics, Vol. 24, pp , 1993.
Markov Random Fields Probabilistic Models for Images
Fast Parallel and Adaptive Updates for Dual-Decomposition Solvers Ozgur Sumer, U. Chicago Umut Acar, MPI-SWS Alexander Ihler, UC Irvine Ramgopal Mettu,
Fast Simulators for Assessment and Propagation of Model Uncertainty* Jim Berger, M.J. Bayarri, German Molina June 20, 2001 SAMO 2001, Madrid *Project of.
Tokyo Institute of Technology, Japan Yu Nishiyama and Sumio Watanabe Theoretical Analysis of Accuracy of Gaussian Belief Propagation.
Readings: K&F: 11.3, 11.5 Yedidia et al. paper from the class website
Mobile Robot Localization (ch. 7)
Daphne Koller Message Passing Belief Propagation Algorithm Probabilistic Graphical Models Inference.
CS Statistical Machine learning Lecture 24
1 Mean Field and Variational Methods finishing off Graphical Models – Carlos Guestrin Carnegie Mellon University November 5 th, 2008 Readings: K&F:
QUIZ!!  In HMMs...  T/F:... the emissions are hidden. FALSE  T/F:... observations are independent given no evidence. FALSE  T/F:... each variable X.
Lecture 2: Statistical learning primer for biologists
by Ryan P. Adams, Iain Murray, and David J.C. MacKay (ICML 2009)
Short Introduction to Particle Filtering by Arthur Pece [ follows my Introduction to Kalman filtering ]
Tracking with dynamics
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Introduction to Sampling Methods Qi Zhao Oct.27,2004.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
Rao-Blackwellised Particle Filtering for Dynamic Bayesian Network Arnaud Doucet Nando de Freitas Kevin Murphy Stuart Russell.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Markov Networks: Theory and Applications Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208
Perfect recall: Every decision node observes all earlier decision nodes and their parents (along a “temporal” order) Sum-max-sum rule (dynamical programming):
Institute of Statistics and Decision Sciences In Defense of a Dissertation Submitted for the Degree of Doctor of Philosophy 26 July 2005 Regression Model.
A Collapsed Variational Bayesian Inference Algorithm for Latent Dirichlet Allocation Yee W. Teh, David Newman and Max Welling Published on NIPS 2006 Discussion.
Probabilistic Robotics
Markov Networks.
Graduate School of Information Sciences, Tohoku University
≠ Particle-based Variational Inference for Continuous Systems
Expectation-Maximization & Belief Propagation
Readings: K&F: 11.3, 11.5 Yedidia et al. paper from the class website
Markov Networks.
Presentation transcript:

Continuous Variables Write message update equation as an expectation: Proposal distribution W t (x t ) for each node Samples define a random discretization of the state space Messages are weightings defined on this discrete domain Samples x t (j), proposal W t held fixed for analysis Particle Belief Propagation Alexander Ihler David McAllester Toyota Technological Institute, Chicago Graphical Models Bren School of Information & Computer Science University of California, Irvine Distribution written in terms of potential functions: Graph Separation Conditional Independence I. Message Product: Multiply incoming messages (from all nodes but s ) with the local observation to form a distribution over x t, II. Message Propagation: Transform distribution from node t to node s using the pairwise potential  s,t Integrate over x t to form f’n summarizing t’s knowledge about x s. Particle Belief PropagationConsistency Convergence Rates Experimental Evaluation Theorem 1: For a tree with k nodes, if we sample n particles at each node with n > k 2 R W ln(kn/  ), and compute the message values defined by (1), then with probability at least 1-  over the choice of particles we have, simultaneously for all nodes s and all particles x s (i), Theorem 2: Under the same conditions as Theorem 1, with probability at least 1-  ’ over the choice of particles we have for all nodes s that Stereo depth maps (Use pairwise; can generalize to higher order) Goal: infer marginal distributions (can use to construct estimators, etc.) Belief Propagation Neighborhood (nodes adjacent to s) Message (represents information from x t about x s ) Belief (approximate marginal) PBP Outline: “Stripped down” particle approximation algorithm Enables theoretical analysis Consistent; n -1/2 convergence rate Related to PAC bounds for learning Convergence results for particle filtering Suggests Proposal distribution choice MCMC particle update procedure Other extensions … in Nonparametric Belief Propagation (NBP) Nodes u, s both send messages to t Samples drawn at source (u,s) to represent messages Sample sets will not overlap! Product = ? Solution: smooth the messages with a Gaussian kernel Leads to sampling from product of Gaussian mixtures Nominally, O(n^d) where d = # of neighbors. …in PBP, Sample locations for incoming messages drawn at destination, t Ensures message particles overlap No smoothing required Product is O(n); Propagation O(n 2 ) Easy to show this algorithm is consistent (approaches true BP messages as n ! 1) Assume large but finite set of states Define states taken on by some particle, and counts Then, messages can be written as and since c t ! W t, we have consistency. Resampling Methods Most stochastic versions of BP have a resampling operation Allows “better” particles to be selected with more information Ex: NBP: smooth messages, draw from product each iteration Koller et al. (1999): Fit distribution, re-draw samples PBP: Can rewrite Suggests drawing samples from the belief B(x t ) (Done in practice by other algorithms previously also) Can use MCMC to sample from the current belief: Use Metropolis-Hastings algorithm Evaluate the ratio of beliefs at any two points via (2) Define a rate constant Analyze finite trees; extensible to more general cases i.e., with high probability our beliefs are accurate at the sample locations. Moreover, we can define a belief estimate at any value of x s by, i.e., with high probability our beliefs are also accurate in an L-1 sense. (2) (1) + (Left image) (Right image) (Depth map) (Graph) From Middlebury data set; Scharstein & Szeliski 2002 Use to evaluate convergence properties: univariate x t, discretization tractable Compare particle approximations to exact Local: W t = local likelihood function True belief: W t = B t (x t ) Estimated belief: use MCMC to sample from current estimate NBP, message & belief-based samples PBP improves at rate n -1/2 as predicted NBP improves at a similar rate slightly slower (~ n -2/5 )? consistent with kernel variance rate (Smoothing seems to hurt performance here) CDE GFH I J B A C D E G F H I J Sensor localization BP messages: tractable closed form for discrete or jontly Gaussian random variables General continuous problems No closed form Discretization becomes intractiable in as few as 2-3 dimensions Need to use approximations (Koller et al. 1999; Coughlan & Ferreira 2002; Sudderth et al. 2003; Isard 2003, 2008; others…) (Ihler et al. 2005) (Sudderth et al. 2003) y2=rand(1,10);