≠ Particle-based Variational Inference for Continuous Systems

Slides:



Advertisements
Similar presentations
Expectation Propagation in Practice
Advertisements

Part 2: Unsupervised Learning
Algorithms for MAP estimation in Markov Random Fields Vladimir Kolmogorov University College London Tutorial at GDR (Optimisation Discrète, Graph Cuts.
Bayesian Belief Propagation
Mobile Robot Localization and Mapping using the Kalman Filter
Linear Time Methods for Propagating Beliefs Min Convolution, Distance Transforms and Box Sums Daniel Huttenlocher Computer Science Department December,
Exact Inference. Inference Basic task for inference: – Compute a posterior distribution for some query variables given some observed evidence – Sum out.
Variational Methods for Graphical Models Micheal I. Jordan Zoubin Ghahramani Tommi S. Jaakkola Lawrence K. Saul Presented by: Afsaneh Shirazi.
ICCV 2007 tutorial Part III Message-passing algorithms for energy minimization Vladimir Kolmogorov University College London.
Dynamic Bayesian Networks (DBNs)
Loopy Belief Propagation a summary. What is inference? Given: –Observabled variables Y –Hidden variables X –Some model of P(X,Y) We want to make some.
Exploiting Sparse Markov and Covariance Structure in Multiresolution Models Presenter: Zhe Chen ECE / CMR Tennessee Technological University October 22,
Convergent Message-Passing Algorithms for Inference over General Graphs with Convex Free Energies Tamir Hazan, Amnon Shashua School of Computer Science.
Variational Inference in Bayesian Submodular Models
Introduction to Belief Propagation and its Generalizations. Max Welling Donald Bren School of Information and Computer and Science University of California.
Parameterized Timing Analysis with General Delay Models and Arbitrary Variation Sources Khaled R. Heloue and Farid N. Najm University of Toronto {khaled,
CS774. Markov Random Field : Theory and Application Lecture 04 Kyomin Jung KAIST Sep
BAYESIAN INFERENCE Sampling techniques
Belief Propagation in a Continuous World Andrew Frank 11/02/2009 Joint work with Alex Ihler and Padhraic Smyth TexPoint fonts used in EMF. Read the TexPoint.
Stochastic Collapsed Variational Bayesian Inference for Latent Dirichlet Allocation James Foulds 1, Levi Boyles 1, Christopher DuBois 2 Padhraic Smyth.
1 Graphical Models in Data Assimilation Problems Alexander Ihler UC Irvine Collaborators: Sergey Kirshner Andrew Robertson Padhraic Smyth.
Extending Expectation Propagation for Graphical Models Yuan (Alan) Qi Joint work with Tom Minka.
Genome evolution: a sequence-centric approach Lecture 5: Undirected models and variational inference.
Evaluating Hypotheses
Conditional Random Fields
Clustering with Bregman Divergences Arindam Banerjee, Srujana Merugu, Inderjit S. Dhillon, Joydeep Ghosh Presented by Rohit Gupta CSci 8980: Machine Learning.
Bayesian Filtering for Location Estimation D. Fox, J. Hightower, L. Liao, D. Schulz, and G. Borriello Presented by: Honggang Zhang.
Lehrstuhl für Informatik 2 Gabriella Kókai: Maschine Learning 1 Evaluating Hypotheses.
Maria-Florina Balcan A Theoretical Model for Learning from Labeled and Unlabeled Data Maria-Florina Balcan & Avrim Blum Carnegie Mellon University, Computer.
. Expressive Graphical Models in Variational Approximations: Chain-Graphs and Hidden Variables Tal El-Hay & Nir Friedman School of Computer Science & Engineering.
6. Experimental Analysis Visible Boltzmann machine with higher-order potentials: Conditional random field (CRF): Exponential random graph model (ERGM):
CS774. Markov Random Field : Theory and Application Lecture 08 Kyomin Jung KAIST Sep
Target Tracking with Binary Proximity Sensors: Fundamental Limits, Minimal Descriptions, and Algorithms N. Shrivastava, R. Mudumbai, U. Madhow, and S.
Planar Cycle Covering Graphs for inference in MRFS The Typhon Algorithm A New Variational Approach to Ground State Computation in Binary Planar Markov.
Probabilistic Graphical Models
MURI: Integrated Fusion, Performance Prediction, and Sensor Management for Automatic Target Exploitation 1 Dynamic Sensor Resource Management for ATE MURI.
Algorithms for MAP estimation in Markov Random Fields Vladimir Kolmogorov University College London.
Continuous Variables Write message update equation as an expectation: Proposal distribution W t (x t ) for each node Samples define a random discretization.
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
Learning With Bayesian Networks Markus Kalisch ETH Zürich.
Update any set S of nodes simultaneously with step-size We show fixed point update is monotone for · 1/|S| Covering Trees and Lower-bounds on Quadratic.
Approximate Inference: Decomposition Methods with Applications to Computer Vision Kyomin Jung ( KAIST ) Joint work with Pushmeet Kohli (Microsoft Research)
Training Conditional Random Fields using Virtual Evidence Boosting Lin Liao, Tanzeem Choudhury †, Dieter Fox, and Henry Kautz University of Washington.
1 Mean Field and Variational Methods finishing off Graphical Models – Carlos Guestrin Carnegie Mellon University November 5 th, 2008 Readings: K&F:
Distributed cooperation and coordination using the Max-Sum algorithm
Perfect recall: Every decision node observes all earlier decision nodes and their parents (along a “temporal” order) Sum-max-sum rule (dynamical programming):
A Collapsed Variational Bayesian Inference Algorithm for Latent Dirichlet Allocation Yee W. Teh, David Newman and Max Welling Published on NIPS 2006 Discussion.
ICPR2004 (24 July, 2004, Cambridge) 1 Probabilistic image processing based on the Q-Ising model by means of the mean- field method and loopy belief propagation.
Expectation Propagation for Graphical Models Yuan (Alan) Qi Joint work with Tom Minka.
Introduction of BP & TRW-S
Learning Deep Generative Models by Ruslan Salakhutdinov
Extending Expectation Propagation for Graphical Models
Dynamical Statistical Shape Priors for Level Set Based Tracking
Outlier Discovery/Anomaly Detection
Approximation of the Propagator for the Harmonic Oscillator
Statistical Methods For Engineers
Markov Networks.
Generalized Belief Propagation
Four-Cut: An Approximate Sampling Procedure for Election Audits
Bayesian Statistics and Belief Networks
Bucket Renormalization for Approximate Inference
Markov Random Fields Presented by: Vladan Radosavljevic.
Algorithms and Theory of
Expectation-Maximization & Belief Propagation
Extending Expectation Propagation for Graphical Models
Approximate Inference by Sampling
Markov Networks.
BP in Practice Message Passing Inference Probabilistic Graphical
Mean Field and Variational Methods Loopy Belief Propagation
Sampling Plans.
Presentation transcript:

≠ Particle-based Variational Inference for Continuous Systems Alexander Ihler Andrew Frank Padhraic Smyth Department of Computer Science, University of California, Irvine Summary Experimental Results (contd.) (2) Importance-reweighted discrete inference f(xs, xt) (1) (3) Xs Xt Sample proposals. (3) Adjust proposals. Recent advances in inference for discrete systems have led to algorithms that offer: Guaranteed convergence Provably more accurate results Bounds on the partition function. We extend these techniques to work in continuous systems by incorporating them into the particle belief propagation framework. Continuous grid marginal estimation MF-PBP Marginal PBP Marginal TRW-PBP Marginal Actual Estimated Weak pairwise potentials: Extensions of Particle BP Handling Continuity Tree-reweighted PBP2: Discretization: Strong pairwise potentials: x y Bounding the partition function “Inclusive” marginal estimation. Convex cost function. Provides an upper bound on the partition function. Works well for some problems, but…. Mean field PBP always converges to a lower bound. Tree-reweighted PBP converges to an upper bound with enough particles. With fewer than 30 particles, the “upper bound” may be below the true value. Variance decreases as the number of particles increases. Very fine discretizations may be required to achieve desired accuracy. Discrete domain size is exponential in the number of continuous dimensions. Mean-field PBP: Parametric Assumptions: Some methods restrict beliefs to simple parametric forms (Gaussian BP, expectation propagation). “Exclusive” marginal estimation. Non-convex cost function. Provides a lower bound on the partition function. Effective when true beliefs match the parametric form. Beliefs may be hard to capture parametrically. Model 2:Sensor Localization: Sensors have noisy pairwise distance measurements of each other. “Anchor” nodes have known locations. Task: compute a marginal distribution over the “target” node’s location. ≠ Many possible extensions: Fractional BP (a generalization of tree-reweighted BP). Expectation propagation (minimizes a “local” alpha divergence). Any other message-passing style discrete inference algorithm… Choose an algorithm based on desired performance characteristics. Exact PBP Run 2 PBP Run 1 TRW-PBP True belief Gaussian approximation Particle Belief Propagation1: Experimental Results A non-parametric adaptive discritization approach. Algorithm overview: The true marginals show two highly probably regions. PBP finds each of them in separate runs, but never both together. TRW-PBP finds both regions, but is less certain. Model 1: Continuous Attractive Grid: Draw samples from proposal distributions over each variable’s domain. Run importance-reweighted belief propagation. Adjust proposals according to partial inference results. Repeat. Pairwise Potentials Local Potentials This toy problem is the continuous analog to the Ising grid. Message update equation: References A. Ihler and D. McAllester. Particle belief propagation. In AI & Statistics: JMLR W&CP, volume 5, pages 256–263, April 2009. M. Wainwright, T. Jaakkola, and A. Willsky. A new class of upper bounds on the log partition function. IEEE Trans. Info. Theory, 51(7):2313–2335, July 2005. 1 ||xs – xt|| xs 1