Belief networks Conditional independence Syntax and semantics Exact inference Approximate inference CS 460, Belief Networks1 Mundhenk and Itti 2008. Based.

Slides:



Advertisements
Similar presentations
Bayesian networks Chapter 14 Section 1 – 2. Outline Syntax Semantics Exact computation.
Advertisements

Probabilistic Reasoning Bayesian Belief Networks Constructing Bayesian Networks Representing Conditional Distributions Summary.
Reasoning under Uncertainty: Marginalization, Conditional Prob., and Bayes Computer Science cpsc322, Lecture 25 (Textbook Chpt ) Nov, 6, 2013.
1 Bayesian Networks Slides from multiple sources: Weng-Keen Wong, School of Electrical Engineering and Computer Science, Oregon State University.
Identifying Conditional Independencies in Bayes Nets Lecture 4.
Slide 1 Reasoning Under Uncertainty: More on BNets structure and construction Jim Little Nov (Textbook 6.3)
1 22c:145 Artificial Intelligence Bayesian Networks Reading: Ch 14. Russell & Norvig.
Bayesian Networks Chapter 14 Section 1, 2, 4. Bayesian networks A simple, graphical notation for conditional independence assertions and hence for compact.
Reasoning under Uncertainty: Conditional Prob., Bayes and Independence Computer Science cpsc322, Lecture 25 (Textbook Chpt ) March, 17, 2010.
Review: Bayesian learning and inference
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11
Bayesian Networks. Motivation The conditional independence assumption made by naïve Bayes classifiers may seem to rigid, especially for classification.
Bayesian networks Chapter 14 Section 1 – 2.
CSCI 5582 Fall 2006 CSCI 5582 Artificial Intelligence Lecture 14 Jim Martin.
Bayesian Belief Networks
Bayesian Belief Network. The decomposition of large probabilistic domains into weakly connected subsets via conditional independence is one of the most.
1 Bayesian Reasoning Chapter 13 CMSC 471 Adapted from slides by Tim Finin and Marie desJardins.
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
CS 561, Session 29 1 Belief networks Conditional independence Syntax and semantics Exact inference Approximate inference.
Bayesian Networks Russell and Norvig: Chapter 14 CMCS424 Fall 2003 based on material from Jean-Claude Latombe, Daphne Koller and Nir Friedman.
Belief Networks Russell and Norvig: Chapter 15 CS121 – Winter 2002.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
Bayesian Reasoning. Tax Data – Naive Bayes Classify: (_, No, Married, 95K, ?)
1 Midterm Exam Mean: 72.7% Max: % Kernel Density Estimation.
1 Probabilistic Belief States and Bayesian Networks (Where we exploit the sparseness of direct interactions among components of a world) R&N: Chap. 14,
Bayesian Networks Russell and Norvig: Chapter 14 CMCS421 Fall 2006.
Bayesian networks More commonly called graphical models A way to depict conditional independence relationships between random variables A compact specification.
1 CMSC 471 Fall 2002 Class #19 – Monday, November 4.
CSCI 121 Special Topics: Bayesian Network Lecture #1: Reasoning Under Uncertainty.
Bayesian Networks Material used 1 Random variables
Bayesian networks Chapter 14. Outline Syntax Semantics.
Bayesian networks Chapter 14 Section 1 – 2. Bayesian networks A simple, graphical notation for conditional independence assertions and hence for compact.
Reasoning Under Uncertainty: Independence and Inference Jim Little Uncertainty 5 Nov 10, 2014 Textbook §6.3.1, 6.5, 6.5.1,
Probabilistic Belief States and Bayesian Networks (Where we exploit the sparseness of direct interactions among components of a world) R&N: Chap. 14, Sect.
1 CS 391L: Machine Learning: Bayesian Learning: Beyond Naïve Bayes Raymond J. Mooney University of Texas at Austin.
Bayesian networks. Motivation We saw that the full joint probability can be used to answer any question about the domain, but can become intractable as.
1 Chapter 14 Probabilistic Reasoning. 2 Outline Syntax of Bayesian networks Semantics of Bayesian networks Efficient representation of conditional distributions.
2 Syntax of Bayesian networks Semantics of Bayesian networks Efficient representation of conditional distributions Exact inference by enumeration Exact.
Baye’s Rule.
Probabilistic Reasoning ECE457 Applied Artificial Intelligence Spring 2007 Lecture #9.
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
, Patrik Huber.  One of our goals: Evaluation of the posterior p(Z|X)  Exact inference  In practice: often infeasible to evaluate the posterior.
Reasoning under Uncertainty: Conditional Prob., Bayes and Independence Computer Science cpsc322, Lecture 25 (Textbook Chpt ) Nov, 5, 2012.
Probabilistic Reasoning [Ch. 14] Bayes Networks – Part 1 ◦Syntax ◦Semantics ◦Parameterized distributions Inference – Part2 ◦Exact inference by enumeration.
Marginalization & Conditioning Marginalization (summing out): for any sets of variables Y and Z: Conditioning(variant of marginalization):
How do I show that two compound propositions are logically equivalent?
Review: Bayesian inference  A general scenario:  Query variables: X  Evidence (observed) variables and their values: E = e  Unobserved variables: Y.
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
CPSC 7373: Artificial Intelligence Lecture 5: Probabilistic Inference Jiang Bian, Fall 2012 University of Arkansas at Little Rock.
1 Probability FOL fails for a domain due to: –Laziness: too much to list the complete set of rules, too hard to use the enormous rules that result –Theoretical.
Bayes network inference  A general scenario:  Query variables: X  Evidence (observed) variables and their values: E = e  Unobserved variables: Y 
Conditional Independence As with absolute independence, the equivalent forms of X and Y being conditionally independent given Z can also be used: P(X|Y,
Belief Networks CS121 – Winter Other Names Bayesian networks Probabilistic networks Causal networks.
PROBABILISTIC REASONING Heng Ji 04/05, 04/08, 2016.
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016.
Web-Mining Agents Data Mining Prof. Dr. Ralf Möller Universität zu Lübeck Institut für Informationssysteme Karsten Martiny (Übungen)
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11 CS479/679 Pattern Recognition Dr. George Bebis.
A Brief Introduction to Bayesian networks
CS 2750: Machine Learning Directed Graphical Models
Bayesian Networks Chapter 14 Section 1, 2, 4.
Bayesian networks Chapter 14 Section 1 – 2.
Presented By S.Yamuna AP/CSE
Probabilistic Reasoning; Network-based reasoning
CS 188: Artificial Intelligence Fall 2008
Class #16 – Tuesday, October 26
Belief Networks CS121 – Winter 2003 Belief Networks.
Bayesian networks Chapter 14 Section 1 – 2.
Probabilistic Reasoning
Presentation transcript:

Belief networks Conditional independence Syntax and semantics Exact inference Approximate inference CS 460, Belief Networks1 Mundhenk and Itti Based on material from S Russel and P Norvig

Independence CS 460, Belief Networks2

Conditional independence CS 460, Belief Networks3

Conditional Independence CS 460, Belief Networks4 Cavity Toothache Catch Other interactions may exist, but they are either insignificant, unknown or irrelevant. We leave them out.

Conditional Independence CS 460, Belief Networks5 Cavity Toothache Catch We assume that a “catch” is not influenced by a toothache and visa versa.

Conditional independence CS 460, Belief Networks6

Conditional Independence CS 460, Belief Networks7 Cavity Toothache Catch (1a) Since Catch does not affect Toothache P(Toothache|Catch,Cavity) = P(Toothache|Cavity) Cavity Toothache =

Conditional Independence CS 460, Belief Networks8 Cavity Toothache Catch (1b) Algebraically these statements are equivalent P(Toothache,Catch|Cavity) = P(Toothache|Cavity)P(Catch|Cavity) Cavity Toothache Catch Cavity =

Conditional independence CS 460, Belief Networks9

Belief networks CS 460, Belief Networks10

CS 460, Belief Networks11 Example Probabilities derived from prior observations

Typical Bayesian Network CS 460, Belief Networks12 Burglary Earthquake Alarm JohnCalls MaryCalls P(B) P(E) BEP(A) TT0.95 TF0.94 FT0.29 FF0.001 AP(J) T0.90 F0.05 AP(M) T0.70 F0.01 Here we see both the Topology and the Conditional Probability Tables (CPT).

CS 460, Belief Networks13 Semantics

Semantics CS 460, Belief Networks14

CS 460, Belief Networks15 Markov blanket

Constructing belief networks CS 460, Belief Networks16

CS 460, Belief Networks17 Burglary Earthquake Alarm JohnCalls MaryCalls P(B) P(E) BEP(A) TT0.95 TF0.94 FT0.29 FF0.001 AP(J) T0.90 F0.05 AP(M) T0.70 F0.01 What is the probability that the alarm has sounded but neither a burglary nor earthquake has occurred and both John and Mary call? – Example: Full Joint Distribution

CS 460, Belief Networks18 Burglary Earthquake Alarm JohnCalls MaryCalls P(B) P(E) BEP(A) TT0.95 TF0.94 FT0.29 FF0.001 AP(J) T0.90 F0.05 AP(M) T0.70 F0.01

What if we find a new variable? CS 460, Belief Networks19 Burglary Earthquake Alarm JohnCalls MaryCalls P(B) P(E) BSEP(A) TTT0.98 TTF0.94 TFT0.96 TFF0.95 FTT0.45 FTF0.25 FFT0.29 FFF0.001 AP(J) T0.90 F0.05 AP(M) T0.70 F0.01 We find that storms can also set off alarms. We add that into our CPT. Notice that JohnCalls and MaryCalls stay the same since Storms were always there but were just unaccounted for. John and Mary did not change! However, we have better precision at P(A). Storm P(S) 0.1

What if we cause a new variable? (or a new variable just occurs) CS 460, Belief Networks20 Burglary Earthquake Alarm JohnCalls MaryCalls P(B) P(E) BCEP(A) TTT0.98 TTF0.94 TFT0.96 TFF0.95 FTT0.45 FTF0.25 FFT0.29 FFF0.001 AP(J) T0.90 F0.05 AP(M) T0.70 F0.01 What if we inject a new cause that was not there before. We pay a crazy guy to set off the alarm frequently, JohnCalls and MaryCalls may no longer be valid since we may have changed the behaviors. For instance the alarm goes off so often now that John and Mary are more likely to ignore it. CrazyGuy P(C) 0.25

What if we cause a new variable? CS 460, Belief Networks21 Burglary Earthquake Alarm JohnCalls MaryCalls P(B) P(E) BCEP(A) TTT0.98 TTF0.94 TFT0.96 TFF0.95 FTT0.45 FTF0.25 FFT0.29 FFF0.001 AP(J) T0.90 F0.05 AP(M) T0.70 F0.01 If the introduced variable is highly erratic, it can invalidate even more of the CPT than we would like. CrazyGuy P(C) 0.25

What if we cause a new variable? CS 460, Belief Networks22 Burglary Earthquake Alarm JohnCalls MaryCalls P(B) P(E) BCEP(A) TTT0.98 TTF0.94 TFT0.96 TFF0.95 FTT0.45 FTF0.25 FFT0.29 FFF0.001 AP(J) T0.90 F0.05 AP(M) T0.70 F0.01 However some changes to the CPT may be absurd so we may never have to worry about them. CrazyGuy P(C) 0.25

Things in the model can change CS 460, Belief Networks23 We can account for change in the model over time (a more advanced topic). John and Mary may be more or less likely to call at certain times. Cyclical repetition may be not too difficult to model. People may become tired of their job and be less likely call over longer periods. This may be easy or difficult to model. Crime picks up. If the trend is slow enough, the model may be able to adjust online even if we have never observed crime picking up before. However, this may easily and totally throw our model off. However, keep in mind, we may get good enough results for our model even without accounting for changes over time.

How would we apply this to Robotics? CS 460, Belief Networks24 Tree in Path Rock in Path Obstacle Detect Execute Stop Execute Turn Execute Turn P(T) 0.25 P(R) 0.4 TRP(O) TT0.95 TF0.45 FT0.29 FF0.001 OP(S) T0.90 F0.05 OP(U) T0.70 F0.01 The CPT can describe other events and probabilities such as action success given observations.

Exact Inference in a Bayesian Network CS 460, Belief Networks25 What if we want to make an inference such as: what is the probability of a tree in the path given that the robot has stopped and turned. This might be useful to a robot which can judge if there is a tree in a path based on the behavior of another robot. So if robot A sees robot B turn or stop it might infer that there is a tree in the path. Normalized we get:

Bayesian Inference cont’ CS 460, Belief Networks26 We compute P of tree and P of not tree and normalize. This is essentially an enumeration of all situations for tree and not tree.

Bayesian Inference cont’ CS 460, Belief Networks27 Finishing: The P of a tree in the path is

CS 460, Belief Networks28 What about hidden variables?

CS 460, Belief Networks29 Example: car diagnosis

CS 460, Belief Networks30 Example: car insurance

Compact conditional distributions CS 460, Belief Networks31

CS 460, Belief Networks32 Compact conditional distributions Know Infer

CS 460, Belief Networks33 Hybrid (discrete+continuous) networks If subsidy was a hidden variable would it be discrete and discreet?

Continuous child variables CS 460, Belief Networks34

Continuous child variables CS 460, Belief Networks35

Discrete variable w/ continuous parents CS 460, Belief Networks36

CS 460, Belief Networks37 Discrete variable

Inference in belief networks Exact inference by enumeration Exact inference by variable elimination Approximate inference by stochastic simulation Approximate inference by Markov chain Monte Carlo (MCMC) CS 460, Belief Networks38