Belief Propagation by Jakob Metzler. Outline Motivation Pearl’s BP Algorithm Turbo Codes Generalized Belief Propagation Free Energies.

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

Exact Inference. Inference Basic task for inference: – Compute a posterior distribution for some query variables given some observed evidence – Sum out.
Exact Inference in Bayes Nets
Junction Trees And Belief Propagation. Junction Trees: Motivation What if we want to compute all marginals, not just one? Doing variable elimination for.
Loopy Belief Propagation a summary. What is inference? Given: –Observabled variables Y –Hidden variables X –Some model of P(X,Y) We want to make some.
For Monday Read chapter 18, sections 1-2 Homework: –Chapter 14, exercise 8 a-d.
An Introduction to Variational Methods for Graphical Models.
Introduction to Belief Propagation and its Generalizations. Max Welling Donald Bren School of Information and Computer and Science University of California.
Pearl’s Belief Propagation Algorithm Exact answers from tree-structured Bayesian networks Heavily based on slides by: Tomas Singliar,
Overview of Inference Algorithms for Bayesian Networks Wei Sun, PhD Assistant Research Professor SEOR Dept. & C4I Center George Mason University, 2009.
Graphical models: approximate inference and learning CA6b, lecture 5.
Belief Propagation on Markov Random Fields Aggeliki Tsoli.
CS774. Markov Random Field : Theory and Application Lecture 04 Kyomin Jung KAIST Sep
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
M.I. Jaime Alfonso Reyes ´Cortés.  The basic task for any probabilistic inference system is to compute the posterior probability distribution for a set.
Bayesian network inference
1 Graphical Models in Data Assimilation Problems Alexander Ihler UC Irvine Collaborators: Sergey Kirshner Andrew Robertson Padhraic Smyth.
Global Approximate Inference Eran Segal Weizmann Institute.
Conditional Random Fields
Propagation in Poly Trees Given a Bayesian Network BN = {G, JDP} JDP(a,b,c,d,e) = p(a)*p(b|a)*p(c|e,b)*p(d)*p(e|d) a d b e c.
Belief Propagation, Junction Trees, and Factor Graphs
Graphical Models Lei Tang. Review of Graphical Models Directed Graph (DAG, Bayesian Network, Belief Network) Typically used to represent causal relationship.
Understanding Belief Propagation and its Applications Dan Yuan June 2004.
© 2005, it - instituto de telecomunicações. Todos os direitos reservados. Gerhard Maierbacher Scalable Coding Solutions for Wireless Sensor Networks IT.
Code and Decoder Design of LDPC Codes for Gbps Systems Jeremy Thorpe Presented to: Microsoft Research
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
24 November, 2011National Tsin Hua University, Taiwan1 Mathematical Structures of Belief Propagation Algorithms in Probabilistic Information Processing.
Computer vision: models, learning and inference
Some Surprises in the Theory of Generalized Belief Propagation Jonathan Yedidia Mitsubishi Electric Research Labs (MERL) Collaborators: Bill Freeman (MIT)
Physics Fluctuomatics (Tohoku University) 1 Physical Fluctuomatics 7th~10th Belief propagation Appendix Kazuyuki Tanaka Graduate School of Information.
Message-Passing for Wireless Scheduling: an Experimental Study Paolo Giaccone (Politecnico di Torino) Devavrat Shah (MIT) ICCCN 2010 – Zurich August 2.
1 Structured Region Graphs: Morphing EP into GBP Max Welling Tom Minka Yee Whye Teh.
For Wednesday Read Chapter 11, sections 1-2 Program 2 due.
Direct Message Passing for Hybrid Bayesian Networks Wei Sun, PhD Assistant Research Professor SFL, C4I Center, SEOR Dept. George Mason University, 2009.
Physics Fluctuomatics / Applied Stochastic Process (Tohoku University) 1 Physical Fluctuomatics Applied Stochastic Process 9th Belief propagation Kazuyuki.
Readings: K&F: 11.3, 11.5 Yedidia et al. paper from the class website
Physics Fluctuomatics (Tohoku University) 1 Physical Fluctuomatics 7th~10th Belief propagation Kazuyuki Tanaka Graduate School of Information Sciences,
1 CMSC 671 Fall 2001 Class #21 – Tuesday, November 13.
CS Statistical Machine learning Lecture 24
Approximate Inference: Decomposition Methods with Applications to Computer Vision Kyomin Jung ( KAIST ) Joint work with Pushmeet Kohli (Microsoft Research)
1 Mean Field and Variational Methods finishing off Graphical Models – Carlos Guestrin Carnegie Mellon University November 5 th, 2008 Readings: K&F:
Lecture 2: Statistical learning primer for biologists
Belief Propagation and its Generalizations Shane Oldenburger.
Wei Sun and KC Chang George Mason University March 2008 Convergence Study of Message Passing In Arbitrary Continuous Bayesian.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
Learning and Acting with Bayes Nets Chapter 20.. Page 2 === A Network and a Training Data.
Graduate School of Information Sciences, Tohoku University
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Daphne Koller Overview Conditional Probability Queries Probabilistic Graphical Models Inference.
Today Graphical Models Representing conditional dependence graphically
Markov Networks: Theory and Applications Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208
Bayesian Belief Propagation for Image Understanding David Rosenberg.
10 October, 2007 University of Glasgow 1 EM Algorithm with Markov Chain Monte Carlo Method for Bayesian Image Analysis Kazuyuki Tanaka Graduate School.
Today.
Factor Graphs and the Sum-Product Algorithm
Exact Inference Continued
Graduate School of Information Sciences, Tohoku University, Japan
Generalized Belief Propagation
Exact Inference ..
Class #19 – Tuesday, November 3
Graduate School of Information Sciences, Tohoku University
Physical Fluctuomatics 7th~10th Belief propagation
Expectation-Maximization & Belief Propagation
Class #16 – Tuesday, October 26
Probabilistic image processing and Bayesian network
Readings: K&F: 11.3, 11.5 Yedidia et al. paper from the class website
Mean Field and Variational Methods Loopy Belief Propagation
Generalized Belief Propagation
Presentation transcript:

Belief Propagation by Jakob Metzler

Outline Motivation Pearl’s BP Algorithm Turbo Codes Generalized Belief Propagation Free Energies

Probabilistic Inference From the lecture we know: Computing the a posteriori belief of a variable in a general Bayesian Network is NP-hard Solution: approximate inference MCMC sampling

Probabilistic Inference From the lecture we know: Computing the a posteriori belief of a variable in a general Bayesian Network is NP-hard Solution: approximate inference MCMC sampling Belief Propagation

In BBNs, we can define the belief BEL(x) of a node x in a graph in the following way: In BP, the pi and lambda terms are messages sent to the node x from its parents and children, respectively BEL(x) = P(x|e) = P(x|e +,e - ) = P(e - |x, e + ) * P(x|e + ) / P(e - |e + )

Pearl’s BP Algorithm Initialization For nodes with evidence e (x i ) = 1 wherever x i = e i ; 0 otherwise  (x i ) = 1 wherever x i = e i ; 0 otherwise For nodes without parents  (x i ) = p(x i ) - prior probabilities For nodes without children (x i ) = 1 uniformly (normalize at end)

Pearl’s BP Algorithm 1. Combine incoming messages from all parents U={U 1,…,U n } into 2. Combine incoming messages from all children Y={Y 1,…Y n } into 3. Compute 4. Send messages to children Y 5. Send messages to parents X

Pearl’s BP Algorithm X U2U2 Y1Y1 Y2Y2 U1U1

X U2U2 Y1Y1 Y2Y2 U1U1

X U2U2 Y1Y1 Y2Y2 U1U1

X U2U2 Y1Y1 Y2Y2 U1U1

X U2U2 Y1Y1 Y2Y2 U1U1

Example of BP in a tree

Properties of BP Exact for polytrees Each node separates Graph into 2 disjoint components On a polytree, the BP algorithm converges in time proportional to diameter of network – at most linear Work done in a node is proportional to the size of CPT Hence BP is linear in number of network parameters For general BBNs Exact inference is NP-hard Approximate inference is NP-hard

Properties of BP Another example of exact inference: Hidden Markov chains Applying BP to its BBN representation yields the forward-backward algorithm

Loopy Belief Propagation Most graphs are not polytrees Cutset conditioning Clustering Join Tree Method Approximate Inference Loopy BP

Loopy Belief Propagation If BP is used on graphs with loops, messages may circulate indefinitely Empirically, a good approximation is still achievable Stop after fixed # of iterations Stop when no significant change in beliefs If solution is not oscillatory but converges, it usually is a good approximation Example: Turbo Codes

Outline Motivation Pearl’s BP Algorithm Turbo Codes Generalized Belief Propagation Free Energies

Decoding Algorithms Information U is to be transmitted reliably over a noisy, memoryless channel U is encoded systematically into codeword X=(U, X 1 ) X 1 is a “codeword fragment It is received as Y=(Y s, Y 1 ) The noisy channel has transition probabilities defined by p(y|x) = Pr{Y=y|X=x}

Decoding Algorithms Since it is also memoryless, we have The decoding problem: Infer U from observed values Y by maximizing the belief: If we define the belief is given by

Decoding using BP We can also represent the problem with a bayesian network: Using BP on this graph is another way of deriving the solution mentioned earlier.

Turbo Codes The information U can also be encoded using 2 encoders: Motivation: using 2 simple encoders in parallel can produce a very effective overall encoding Interleaver causes permutation of inputs

Turbo Codes The bayesian network corresponding to the decoding problem is not a polytree any more but has loops:

Turbo Codes We can still approximate the optimal beliefs by using loopy Belief Propagation Stop after # of iterations Choosing the order of belief updates among the nodes derives “different” previously known algorithms Sequence U->X 1 ->U->X 2 ->U->X 1 ->U etc. yields the well- known turbo decoding algorithm Sequence U->X->U->X->U etc. yields a general decoding algorithm for multiple turbo codes and many more

Turbo Codes Summary BP can be used as a general decoding algorithm by representing the problem as a BBN and running BP on it. Many existing, seemingly different decoding algorithms are just instantiations of BP. Turbo codes are a good example of successful convergence of BP on a loopy graph.

Outline Motivation Pearl’s BP Algorithm Turbo Codes Generalized Belief Propagation Free Energies

BP in MRFs BP can also be applied to other graphical models, e.g. pairwise MRFs Hidden variables x i and x j are connected through a compatibility function Hidden variables x i are connected to observable variables y i by the local “evidence” function Pairwise, so it can also be abbreviated as The joint probability of {x} is given by

BP in MRFs In pairwise MRFs, the messages & beliefs are updated the following way:

Example

Generalized BP We can try to improve inference by taking into account higher-order interactions among the variables An intuitive way to do this is to define messages that propagate between groups of nodes rather than just single nodes This is the intuition in Generalized Belief Propagation (GPB)

GBP Algorithm 1) Split the graph into basic clusters [1245],[2356], [4578],[5689]

GBP Algorithm 2) Find all intersection regions of the basic clusters, and all their intersections [25], [45], [56], [58], [5]

GBP Algorithm 3) Create a hierarchy of regions and their direct sub- regions

GBP Algorithm 4) Associate a message with each line in the graph e.g. message from [1245]->[25]: m 14->25 (x 2,x 5 )

GBP Algorithm 5) Setup equations for beliefs of regions - remember from earlier: - So the belief for the region containing [5] is: - for the region [45]: - etc.

GBP Algorithm 6) Setup equations for updating messages by enforcing marginalization conditions and combining them with the belief equations: e.g. condition yields, with the previous two belief formulas, the message update rule

Experiment [Yedidia et al., 2000]: “square lattice Ising spin glass in a random magnetic field” Structure: Nodes are arranged in square lattice of size n*n Compatibility matrix: Evidence term:

Experiment Results For n>=20, ordinary BP did not converge For n=10: (marginals) BPGBPExact

Outline Motivation Pearl’s BP Algorithm Turbo Codes Generalized Belief Propagation Free Energies