Conditional Independence As with absolute independence, the equivalent forms of X and Y being conditionally independent given Z can also be used: P(X|Y,

Slides:



Advertisements
Similar presentations
Bayesian networks Chapter 14 Section 1 – 2. Outline Syntax Semantics Exact computation.
Advertisements

Bayesian Networks CSE 473. © Daniel S. Weld 2 Last Time Basic notions Atomic events Probabilities Joint distribution Inference by enumeration Independence.
1 Knowledge Engineering for Bayesian Networks. 2 Probability theory for representing uncertainty l Assigns a numerical degree of belief between 0 and.
Identifying Conditional Independencies in Bayes Nets Lecture 4.
For Monday Read chapter 18, sections 1-2 Homework: –Chapter 14, exercise 8 a-d.
For Monday Finish chapter 14 Homework: –Chapter 13, exercises 8, 15.
Bayesian Networks Chapter 14 Section 1, 2, 4. Bayesian networks A simple, graphical notation for conditional independence assertions and hence for compact.
CPSC 322, Lecture 26Slide 1 Reasoning Under Uncertainty: Belief Networks Computer Science cpsc322, Lecture 27 (Textbook Chpt 6.3) March, 16, 2009.
3/19. Conditional Independence Assertions We write X || Y | Z to say that the set of variables X is conditionally independent of the set of variables.
Review: Bayesian learning and inference
Bayesian Networks. Motivation The conditional independence assumption made by naïve Bayes classifiers may seem to rigid, especially for classification.
Probabilistic Reasoning Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 14 (14.1, 14.2, 14.3, 14.4) Capturing uncertain knowledge Probabilistic.
Bayesian networks Chapter 14 Section 1 – 2.
Bayesian Belief Networks
Bayesian Networks. Graphical Models Bayesian networks Conditional random fields etc.
1 Bayesian Reasoning Chapter 13 CMSC 471 Adapted from slides by Tim Finin and Marie desJardins.
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
Bayesian Reasoning. Tax Data – Naive Bayes Classify: (_, No, Married, 95K, ?)
. PGM 2002/3 – Tirgul6 Approximate Inference: Sampling.
1 Probabilistic Belief States and Bayesian Networks (Where we exploit the sparseness of direct interactions among components of a world) R&N: Chap. 14,
Bayesian networks More commonly called graphical models A way to depict conditional independence relationships between random variables A compact specification.
1 CMSC 471 Fall 2002 Class #19 – Monday, November 4.
Probabilistic Reasoning
Quiz 4: Mean: 7.0/8.0 (= 88%) Median: 7.5/8.0 (= 94%)
EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS
Artificial Intelligence CS 165A Tuesday, November 27, 2007  Probabilistic Reasoning (Ch 14)
Bayesian networks Chapter 14. Outline Syntax Semantics.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
Artificial Intelligence CS 165A Thursday, November 29, 2007  Probabilistic Reasoning / Bayesian networks (Ch 14)
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
Probabilistic Belief States and Bayesian Networks (Where we exploit the sparseness of direct interactions among components of a world) R&N: Chap. 14, Sect.
1 CS 391L: Machine Learning: Bayesian Learning: Beyond Naïve Bayes Raymond J. Mooney University of Texas at Austin.
Bayesian networks. Motivation We saw that the full joint probability can be used to answer any question about the domain, but can become intractable as.
1 Chapter 14 Probabilistic Reasoning. 2 Outline Syntax of Bayesian networks Semantics of Bayesian networks Efficient representation of conditional distributions.
For Wednesday Read Chapter 11, sections 1-2 Program 2 due.
2 Syntax of Bayesian networks Semantics of Bayesian networks Efficient representation of conditional distributions Exact inference by enumeration Exact.
Made by: Maor Levy, Temple University  Inference in Bayes Nets ◦ What is the probability of getting a strong letter? ◦ We want to compute the.
Announcements Project 4: Ghostbusters Homework 7
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
Marginalization & Conditioning Marginalization (summing out): for any sets of variables Y and Z: Conditioning(variant of marginalization):
Bayesian Networks CSE 473. © D. Weld and D. Fox 2 Bayes Nets In general, joint distribution P over set of variables (X 1 x... x X n ) requires exponential.
Review: Bayesian inference  A general scenario:  Query variables: X  Evidence (observed) variables and their values: E = e  Unobserved variables: Y.
Lecture 29 Conditional Independence, Bayesian networks intro Ch 6.3, 6.3.1, 6.5, 6.5.1,
Inference Algorithms for Bayes Networks
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016.
Web-Mining Agents Data Mining Prof. Dr. Ralf Möller Universität zu Lübeck Institut für Informationssysteme Karsten Martiny (Übungen)
CS 2750: Machine Learning Directed Graphical Models
Bayesian Networks Chapter 14 Section 1, 2, 4.
Bayesian networks Chapter 14 Section 1 – 2.
Qian Liu CSE spring University of Pennsylvania
Inference in Bayesian Networks
Approximate Inference
Bayesian Networks Probability In AI.
CS 4/527: Artificial Intelligence
Structure and Semantics of BN
Instructors: Fei Fang (This Lecture) and Dave Touretzky
CAP 5636 – Advanced Artificial Intelligence
CS 188: Artificial Intelligence Fall 2007
Class #19 – Tuesday, November 3
CS 188: Artificial Intelligence
Class #16 – Tuesday, October 26
Structure and Semantics of BN
Hankz Hankui Zhuo Bayesian Networks Hankz Hankui Zhuo
Belief Networks CS121 – Winter 2003 Belief Networks.
Bayesian Networks CSE 573.
Probabilistic Reasoning
Presentation transcript:

Conditional Independence As with absolute independence, the equivalent forms of X and Y being conditionally independent given Z can also be used: P(X|Y, Z) = P(X|Z) and P(Y|X, Z) = P(Y|Z)

Example of the Key Property The following conditional independence holds: P (MaryCalls |JohnCalls, Alarm, Earthquake, Burglary) = P(MaryCalls | Alarm)

Conditional Independence Again A node X is conditionally independent of its non-descendants given Parents(X). Markov Blanket of X: set consisting of the parents of X, the children of X, and the other parents of the children of X. X is conditionally independent of all other nodes in the network given its Markov Blanket.

Representation Size At first glance it would appear that the space to represent a Bayes Net is necessarily quadratic (possible number of arcs) in the number of random variables. But recall we also must represent the CPT of each node, which in general will have size exponential in the number of parents of the node.

Representation Size If we assume n boolean variables the network can be specified by n2 k numbers in contrast to 2 n for the joint. For example: n=20 and each has at most k = 5, BN requires 640 numbers but the full joint needs over a million!

The Basic Inference Task in Bayesian Networks Given some observed event (some assignment of values to a set of evidence variables), compute the posterior probability distribution over a set of query variables. Variables that are neither evidence variables nor query variables are hidden variables. Most common query: P(X|e).

Generality of Approach A Bayes Net is flexible enough that any variable can be the query variables, and any variables can be evidence variables. Algorithms we consider are easily extended to handle queries with multiple query variables.

Example Query P(Burglary |JohnCalls=true, MaryCalls=true) = How can we compute such answers? One approach is to compute entire full joint distribution represented by the network, and use our earlier algorithm. But this would defeat the entire purpose of Bayes Nets.

Inference By Enumeration Recall the E NUMERATE J OINT A SK procedure based on the preceding equation, which computed the answers to queries given the full joint represented as a table. We will modify E NUMERATE J OINT A SK to use the Bayes Net instead of the table. Need only get desired table entries using Bayes Net.

Getting a Full Joint Table Entry from a Bayes Net Recall: A table entry for X1 = x1,…,Xn = xn is simply P(x1,…,xn) which can be calculated based on the Bayes Net semantics above. Therefore the CPTs provide a decomposed representation of the joint. Recall example:

Example of Full Procedure Query: P(Burglary |JohnCalls=true, MaryCalls=true) P(B|j,m) = Solve without the normalization constant for both B = true and B = false, and then compute the normalization constant and the final probabilities.

Example (Continued)

Normalizing:  ( ) = 1.0  = 1.0/ = 479 P(B|j,m) =