Bucket Elimination: A Unifying Framework for Probabilistic Inference By: Rina Dechter Presented by: Gal Lavee.

Slides:



Advertisements
Similar presentations
Exact Inference. Inference Basic task for inference: – Compute a posterior distribution for some query variables given some observed evidence – Sum out.
Advertisements

Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
. Exact Inference in Bayesian Networks Lecture 9.
“Using Weighted MAX-SAT Engines to Solve MPE” -- by James D. Park Shuo (Olivia) Yang.
Bucket Elimination: A unifying framework for Probabilistic inference Rina Dechter presented by Anton Bezuglov, Hrishikesh Goradia CSCE 582 Fall02 Instructor:
Bayesian Networks Bucket Elimination Algorithm 主講人:虞台文 大同大學資工所 智慧型多媒體研究室.
CSCE 582 Computation of the Most Probable Explanation in Bayesian Networks using Bucket Elimination -Hareesh Lingareddy University of South Carolina.
Constraint Optimization Presentation by Nathan Stender Chapter 13 of Constraint Processing by Rina Dechter 3/25/20131Constraint Optimization.
Exact Inference in Bayes Nets
MPE, MAP AND APPROXIMATIONS Lecture 10: Statistical Methods in AI/ML Vibhav Gogate The University of Texas at Dallas Readings: AD Chapter 10.
A. Darwiche Inference in Bayesian Networks. A. Darwiche Query Types Pr: –Evidence: Pr(e) –Posterior marginals: Pr(x|e) for every X MPE: Most probable.
Ai in game programming it university of copenhagen Statistical Learning Methods Marco Loog.
Junction Trees: Motivation Standard algorithms (e.g., variable elimination) are inefficient if the undirected graph underlying the Bayes Net contains cycles.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
Lecture 5: Learning models using EM
1 Exact Inference Algorithms for Probabilistic Reasoning; COMPSCI 276 Fall 2007.
. Hidden Markov Models For Genetic Linkage Analysis Lecture #4 Prepared by Dan Geiger.
Another example p := &x; *p := 5 y := x + 1;. Another example p := &x; *p := 5 y := x + 1; x := 5; *p := 3 y := x + 1; ???
Tutorial #9 by Ma’ayan Fishelson
Hidden Markov Model Continues …. Finite State Markov Chain A discrete time stochastic process, consisting of a domain D of m states {1,…,m} and 1.An m.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
AND/OR Search for Mixed Networks #CSP Robert Mateescu ICS280 Spring Current Topics in Graphical Models Professor Rina Dechter.
Aspects of Bayesian Inference and Statistical Disclosure Control in Python Duncan Smith Confidentiality and Privacy Group CCSR University of Manchester.
Computer vision: models, learning and inference
. Basic Model For Genetic Linkage Analysis Lecture #5 Prepared by Dan Geiger.
Learning Structure in Bayes Nets (Typically also learn CPTs here) Given the set of random variables (features), the space of all possible networks.
Ahsanul Haque *, Swarup Chandra *, Latifur Khan * and Charu Aggarwal + * Department of Computer Science, University of Texas at Dallas + IBM T. J. Watson.
Ahsanul Haque *, Swarup Chandra *, Latifur Khan * and Michael Baron + * Department of Computer Science, University of Texas at Dallas + Department of Mathematical.
DIFFERENTIATING “COMBINED” FUNCTIONS ---PART I CONSTANT MULTIPLES, SUMS AND DIFFERENCES.
Introduction to Bayesian Networks
1 Variable Elimination Graphical Models – Carlos Guestrin Carnegie Mellon University October 11 th, 2006 Readings: K&F: 8.1, 8.2, 8.3,
Maximum Likelihood - "Frequentist" inference x 1,x 2,....,x n ~ iid N( ,  2 ) Joint pdf for the whole random sample Maximum likelihood estimates.
Made by: Maor Levy, Temple University  Inference in Bayes Nets ◦ What is the probability of getting a strong letter? ◦ We want to compute the.
Probabilistic Networks Chapter 14 of Dechter’s CP textbook Speaker: Daniel Geschwender April 1, 2013 April 1&3, 2013DanielG--Probabilistic Networks1.
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
Probabilistic Graphical Models seminar 15/16 ( ) Haim Kaplan Tel Aviv University.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
1 Chapter 17 2 nd Part Making Complex Decisions --- Decision-theoretic Agent Design Xin Lu 11/04/2002.
1 Tutorial #9 by Ma’ayan Fishelson. 2 Bucket Elimination Algorithm An algorithm for performing inference in a Bayesian network. Similar algorithms can.
1 Relational Factor Graphs Lin Liao Joint work with Dieter Fox.
1 Structure Learning (The Good), The Bad, The Ugly Inference Graphical Models – Carlos Guestrin Carnegie Mellon University October 13 th, 2008 Readings:
Daphne Koller Overview Maximum a posteriori (MAP) Probabilistic Graphical Models Inference.
1 Variable Elimination Graphical Models – Carlos Guestrin Carnegie Mellon University October 15 th, 2008 Readings: K&F: 8.1, 8.2, 8.3,
Machine Learning: Ensemble Methods
Computer Science cpsc322, Lecture 25
Constraint-Based Routing
Maximum Expected Utility
Quizzz Rihanna’s car engine does not start (E).
Markov ó Kalman Filter Localization
Data Mining Lecture 11.
A Distributed Bucket Elimination Algorithm
More about Posterior Distributions
Dynamic Programming General Idea
CSCI 5822 Probabilistic Models of Human and Machine Learning
Instructor: Mohamed Eltabakh
Readings: K&F: 15.1, 15.2, 15.3, 15.4, 15.5 K&F: 7 (overview of inference) K&F: 8.1, 8.2 (Variable Elimination) Structure Learning in BNs 3: (the good,
Reasoning Under Uncertainty: Bnet Inference
Reasoning Under Uncertainty: Bnet Inference
13. Acting under Uncertainty Wolfram Burgard and Bernhard Nebel
Maths Unit 1 - Algebra Order of operations - BIDMAS Algebraic Notation
Markov Random Fields Presented by: Vladan Radosavljevic.
Chapter14-cont..
Dynamic Programming General Idea
Text Categorization Berlin Chen 2003 Reference:
Machine Learning: Lecture 6
Variable Elimination Graphical Models – Carlos Guestrin
Hidden Markov Models ..
presented by Anton Bezuglov, Hrishikesh Goradia CSCE 582 Fall02
Presentation transcript:

Bucket Elimination: A Unifying Framework for Probabilistic Inference By: Rina Dechter Presented by: Gal Lavee

Multiplying Tables Often in inference we see the notation: f(A) * g(B) where A and B are sets of variables and f and g are functions (usually probability functions) given as tables of the form: Example: a function f of two variables: A1 and A2 Where ai and ~ai are the possible states of Ai ~a1a1 f(~a1,a2)f(a1,a2)a2 f(~a1,~a2)f(a1,~a2)~a2

Multiplying Tables Multiply the two functions (given as tables) f and g. Where g takes one variable (B1) and f takes two variables (A1,A2) ~a1a1 f(~a1,a2)f(a1,a2)a2 f(~a1,~a2)f(a1,~a2)~a2 ~b1b1 g(~b1)g(b1)

Multiplying Tables ~a1a1b1 f(~a1,a2)*g(b1)f(a1,a2)*g(b1) a2 f(~a1,~a2)*g(b1)f(a1,~a2)*g(b1) ~a2 ~a1a1~b1 f(~a1,a2)*g(~b1)f(a1,a2)*g(~b1) a2 f(~a1,~a2)*g(~b1)f(a1,~a2)*g(~b1) ~a2 The result is a function h, which takes as 3 arguments (the union of the arguments of f and g), A1,A2,B1 The value of the function h(A1,A2,B1) is simply the product f(A1,A2) * g(B1) This function can be expressed as a table with 2^3 entries:

Summing over tables Given a function of multiple variables it is often useful to eliminate variables by summation Summing the function f of variables A1,A2 (given as a table) over the variable A2 yields a new function f1 of just one variable(A1). The new function value is given by f(A1,a2)+f(A1,~a2). This function can be specified as the table: ~a1a1 f(~a1,a2)f(a1,a2)a2 f(~a1,~a2)f(a1,~a2)~a2 ~a1a1 f(~a1,a2)+ f(~a1,~a2) f(a1,a2)+ f(a1,~a2)

Bucket Elimination-Definition Bucket elimination is a unifying algorithmic framework that generalizes dynamic programming to accommodate algorithms for many complex problem-solving and reasoning activities Including: –Belief Assessment –Most Probable Estimation (MPE) –Maximum A posteriori Hypothesis (MAP) –Maximum Expected Utility MEU Uses “buckets”, an organizational device, to mimic the algebraic manipulations involved in each of these problems resulting in an easily expressible algorithmic formulation Each bucket elimination algorithm is associated with a node ordering

The Bucket Elimination Algorithm Mimics algebraic manipulation Partition functions on the graph into “buckets” in backwards relative to the given node order In the bucket of variable Xi we put all functions that mention Xi but do not mention any variable having a higher index e.g. in the bucketG we will place all functions whose scope includes variable G

The Algorithmcntd. Process buckets backwards relative to the node order Processing a bucket amounts to eliminating the variable in the bucket from subsequent computation. The computed function after elimination is placed in the bucket of the ‘highest’ variable in its scope E.g= sum over G is computed and placed in bucket F

Example - Belief Updating/Assessment Definition –Given a set of evidence compute the posterior probability of each proposition –The belief assessment task of X k = x k is to find where k – normalizing constant

Example - Belief Assessment F B C D G Node Ordering: A,C,B,F,D,G

Example - Belief Assessment F B C D G Node Ordering: A,C,B,F,D,G Initial Bucket Partition

Example - Belief Assessment F B C D G Node Ordering: A,C,B,F,D,G Process buckets in reverse order (G)

Example - Belief Assessment F B C D G Node Ordering: A,C,B,F,D,G Process buckets in reverse order (D)

Example - Belief Assessment F B C D G Node Ordering: A,C,B,F,D,G Process buckets in reverse order (F)

Example - Belief Assessment F B C D G Node Ordering: A,C,B,F,D,G Process buckets in reverse order (B)

Example - Belief Assessment F B C D G Node Ordering: A,C,B,F,D,G Process buckets in reverse order (C)

Example - Belief Assessment F B C D G Node Ordering: A,C,B,F,D,G Process buckets in reverse order (A)

Belief Assessment Summary In reverse Node Ordering: –Create bucket function by multiplying all functions (given as tables) containing the current node –Perform variable elimination using Summation over the current node –Place the new created function table) into the appropriate bucket

Most Probable Explanation (MPE) Definition –Given evidence find the maximum probabillity assignment to the remaining variables –The MPE task is to find an assignment x o = (x o 1, …, x o n ) such that

Most Probable Explanation (MPE) Differences from Belief Assessment –Replace Sums With Maxs –Keep track of maximizing value at each stage –“Forward Step” to determine what is the maximizing assignment tuple

Example - MPE F B C D G Node Ordering: A,C,B,F,D,G

Example - MPE F B C D G Node Ordering: A,C,B,F,D,G Initial Bucket Partition (same as belief assessment)

Example - MPE F B C D G Node Ordering: A,C,B,F,D,G Process buckets in reverse order (G)

Example - MPE F B C D G Node Ordering: A,C,B,F,D,G Process buckets in reverse order (D)

Example - MPE F B C D G Node Ordering: A,C,B,F,D,G Process buckets in reverse order (F)

Example - MPE F B C D G Node Ordering: A,C,B,F,D,G Process buckets in reverse order (B)

Example - MPE F B C D G Node Ordering: A,C,B,F,D,G Process buckets in reverse order (C)

Example - MPE F B C D G Node Ordering: A,C,B,F,D,G Process buckets in reverse order (A)

Example - MPE F B C D G Node Ordering: A,C,B,F,D,G Forward Phase: determine maximizing tuple:

MPE Summary In reverse node Ordering –Create bucket function by multiplying all functions (given as tables) containing the current node –Perform variable elimination using the Maximization operation over the current node (recording the maximizing state function) –Place the new created function table) into the appropriate bucket In forward node ordering –Calculate the maximum probability using maximizing state functions

Maximum Aposteriori Hypothesis (MAP) Definition –Given evidence find an assignment to a subset of “hypothesis” variables that maximizes their probability –Given a set of hypothesis variables A = {A 1, …, A k },, the MAP task is to find an assignment a o = (a o 1, …, a o k ) such that

Maximum Aposteriori Hypothesis (MAP) Mixture of Belief Assessment and MPE –Sum over non-hypothesis vars –Max over hypothesis var –Forward phase over hypothesis vars only –Algorithm presented assumes hypothesis vars are at the beginning of the order

Example - MAP F B C D G Node Ordering: A,C,B,F,D,G Hypothesis Nodes: A,B,C

Example - MAP Initial Bucket Partition (Same as Belief Assessment, MPE) F B C D G Node Ordering: A,C,B,F,D,G Hypothesis Nodes: A,B,C

Example - MAP Process buckets in reverse order (G) F B C D G Node Ordering: A,C,B,F,D,G Hypothesis Nodes: A,B,C

Example - MAP Process buckets in reverse order (D) F B C D G Node Ordering: A,C,B,F,D,G Hypothesis Nodes: A,B,C

Example - MAP Process buckets in reverse order (F) F B C D G Node Ordering: A,C,B,F,D,G Hypothesis Nodes: A,B,C

Example - MAP Process buckets in reverse order (B) F B C D G Node Ordering: A,C,B,F,D,G Hypothesis Nodes: A,B,C

Example - MAP Process buckets in reverse order (C) F B C D G Node Ordering: A,C,B,F,D,G Hypothesis Nodes: A,B,C

Example - MAP Process buckets in reverse order (A) F B C D G Node Ordering: A,C,B,F,D,G Hypothesis Nodes: A,B,C

Example - MAP Forward Phase: determine maximizing tuple: F B C D G Node Ordering: A,C,B,F,D,G Hypothesis Nodes: A,B,C

MAP -Summary In reverse node Ordering –Create bucket function by multiplying all functions (given as tables) containing the current node –If is a hypothesis node: Perform variable elimination using the Maximization operation over the current node (recording the maximizing state function) –Else: Perform variable elimination using Summation over the current node (recording the maximizing state function) –Place the new created function (table) into the appropriate bucket In forward node ordering –Calculate the maximum probability using maximizing state functions over hypothesis nodes only

Generalizing Bucket Elimination As we have seen bucket elimination algorithms can be formulated for different problems by simply altering the elimination operation. For some problems we may also want to change the combination operation which we use to combine functions. Elimination OperatorCombination OperatorProblem SummationMultiplicationBelief Assessment MaximizationMultiplicationMPE MultiplicationJoinCSP MinimizationSummationMax-CSP

Thank You