Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.

Slides:



Advertisements
Similar presentations
Bayesian networks Chapter 14 Section 1 – 2. Outline Syntax Semantics Exact computation.
Advertisements

Probabilistic Reasoning Bayesian Belief Networks Constructing Bayesian Networks Representing Conditional Distributions Summary.
BAYESIAN NETWORKS. Bayesian Network Motivation  We want a representation and reasoning system that is based on conditional independence  Compact yet.
1 Bayesian Networks Slides from multiple sources: Weng-Keen Wong, School of Electrical Engineering and Computer Science, Oregon State University.
Identifying Conditional Independencies in Bayes Nets Lecture 4.
1 22c:145 Artificial Intelligence Bayesian Networks Reading: Ch 14. Russell & Norvig.
Bayesian Networks Chapter 14 Section 1, 2, 4. Bayesian networks A simple, graphical notation for conditional independence assertions and hence for compact.
CPSC 322, Lecture 26Slide 1 Reasoning Under Uncertainty: Belief Networks Computer Science cpsc322, Lecture 27 (Textbook Chpt 6.3) March, 16, 2009.
3/19. Conditional Independence Assertions We write X || Y | Z to say that the set of variables X is conditionally independent of the set of variables.
Review: Bayesian learning and inference
Bayesian Networks. Motivation The conditional independence assumption made by naïve Bayes classifiers may seem to rigid, especially for classification.
Artificial Intelligence Probabilistic reasoning Fall 2008 professor: Luigi Ceccaroni.
Probabilistic Reasoning Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 14 (14.1, 14.2, 14.3, 14.4) Capturing uncertain knowledge Probabilistic.
1 Data Mining with Bayesian Networks (I) Instructor: Qiang Yang Hong Kong University of Science and Technology Thanks: Dan Weld, Eibe.
Bayesian networks Chapter 14 Section 1 – 2.
CSCI 5582 Fall 2006 CSCI 5582 Artificial Intelligence Lecture 14 Jim Martin.
Bayesian Belief Networks
Bayesian Belief Network. The decomposition of large probabilistic domains into weakly connected subsets via conditional independence is one of the most.
University College Cork (Ireland) Department of Civil and Environmental Engineering Course: Engineering Artificial Intelligence Dr. Radu Marinescu Lecture.
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
Bayesian Reasoning. Tax Data – Naive Bayes Classify: (_, No, Married, 95K, ?)
Bayesian networks More commonly called graphical models A way to depict conditional independence relationships between random variables A compact specification.
Probabilistic Reasoning
EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS
Bayesian Networks Material used 1 Random variables
Artificial Intelligence CS 165A Tuesday, November 27, 2007  Probabilistic Reasoning (Ch 14)
Read R&N Ch Next lecture: Read R&N
Bayesian networks Chapter 14. Outline Syntax Semantics.
Bayesian Networks Tamara Berg CS Artificial Intelligence Many slides throughout the course adapted from Svetlana Lazebnik, Dan Klein, Stuart Russell,
Bayesian networks Chapter 14 Section 1 – 2. Bayesian networks A simple, graphical notation for conditional independence assertions and hence for compact.
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
Probabilistic Belief States and Bayesian Networks (Where we exploit the sparseness of direct interactions among components of a world) R&N: Chap. 14, Sect.
Bayesian networks. Motivation We saw that the full joint probability can be used to answer any question about the domain, but can become intractable as.
1 Chapter 14 Probabilistic Reasoning. 2 Outline Syntax of Bayesian networks Semantics of Bayesian networks Efficient representation of conditional distributions.
2 Syntax of Bayesian networks Semantics of Bayesian networks Efficient representation of conditional distributions Exact inference by enumeration Exact.
1 Monte Carlo Artificial Intelligence: Bayesian Networks.
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
Probabilistic Reasoning [Ch. 14] Bayes Networks – Part 1 ◦Syntax ◦Semantics ◦Parameterized distributions Inference – Part2 ◦Exact inference by enumeration.
Marginalization & Conditioning Marginalization (summing out): for any sets of variables Y and Z: Conditioning(variant of marginalization):
Bayesian Networks CSE 473. © D. Weld and D. Fox 2 Bayes Nets In general, joint distribution P over set of variables (X 1 x... x X n ) requires exponential.
Review: Bayesian inference  A general scenario:  Query variables: X  Evidence (observed) variables and their values: E = e  Unobserved variables: Y.
CPSC 7373: Artificial Intelligence Lecture 5: Probabilistic Inference Jiang Bian, Fall 2012 University of Arkansas at Little Rock.
1 Probability FOL fails for a domain due to: –Laziness: too much to list the complete set of rules, too hard to use the enormous rules that result –Theoretical.
Conditional Probability, Bayes’ Theorem, and Belief Networks CISC 2315 Discrete Structures Spring2010 Professor William G. Tanner, Jr.
Belief Networks Kostas Kontogiannis E&CE 457. Belief Networks A belief network is a graph in which the following holds: –A set of random variables makes.
CPSC 322, Lecture 26Slide 1 Reasoning Under Uncertainty: Belief Networks Computer Science cpsc322, Lecture 27 (Textbook Chpt 6.3) Nov, 13, 2013.
Bayesian networks Chapter 14 Slide Set 2. Constructing Bayesian networks 1. Choose an ordering of variables X 1, …,X n 2. For i = 1 to n –add X i to the.
PROBABILISTIC REASONING Heng Ji 04/05, 04/08, 2016.
Web-Mining Agents Data Mining Prof. Dr. Ralf Möller Universität zu Lübeck Institut für Informationssysteme Karsten Martiny (Übungen)
A Brief Introduction to Bayesian networks
Another look at Bayesian inference
Reasoning Under Uncertainty: Belief Networks
Bayesian Networks Chapter 14 Section 1, 2, 4.
Bayesian networks Chapter 14 Section 1 – 2.
Presented By S.Yamuna AP/CSE
Qian Liu CSE spring University of Pennsylvania
Read R&N Ch Next lecture: Read R&N
Conditional Probability, Bayes’ Theorem, and Belief Networks
Read R&N Ch Next lecture: Read R&N
Probabilistic Reasoning; Network-based reasoning
CS 188: Artificial Intelligence Fall 2007
CS 188: Artificial Intelligence Fall 2008
Belief Networks CS121 – Winter 2003 Belief Networks.
Bayesian networks Chapter 14 Section 1 – 2.
Probabilistic Reasoning
Read R&N Ch Next lecture: Read R&N
Chapter 14 February 26, 2004.
Presentation transcript:

Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University

Bayesian Networks2 Course Outline Part I – Introduction to Artificial Intelligence Part II – Classical Artificial Intelligence Part III – Machine Learning Introduction to Machine Learning Neural Networks Probabilistic Reasoning and Bayesian Belief Networks Artificial Life: Learning through Emergent Behavior Part IV – Advanced Topics

Bayesian Networks3 Learning Objectives

Bayesian Networks4 Unit Outline 1. Introduction Introduction 2. Bayesian networks Bayesian networks 3. Summary Summary

Bayesian Networks5 1. Introduction A systematic way, called Bayesian networks, to represent the world probabilistically – independence and conditional independence relationships Application areas: consumer help desks, nuclear reactor diagnosis, tissue pathology, pattern recognition, credit assessment, computer network diagnosis,... Representing knowledge in an uncertain domain Semantics of Bayesian networks Example: Topics

Bayesian Networks6 2. Bayesian networks 1. Definitions Definitions 2. Example – Weather, Cavity, Catch, and Toothache Example – Weather, Cavity, Catch, and Toothache 3. Example – Burglary net Example – Burglary net 4. How to answer queries in the Burglary net How to answer queries in the Burglary net 5. How to construct Bayesian networks How to construct Bayesian networks

Bayesian Networks7 Definitions Bayesian network, belief network, probabilistic network, causal network, or also called knowledge map: A simple, graphical notation for conditional independence assertions and hence for compact specification of full joint distributions Syntax: a set of nodes, one per random variable a directed, acyclic graph (link means “a parent node directly influences its child nodes.”) a conditional probability distribution for each node given its parents: P (X i | Parents (X i )) In the simplest case, conditional probability distribution represented as a conditional probability table (CPT) giving the distribution over X i for each combination of parent values 2.

Bayesian Networks8 Example – Weather, Toothache, C, C Topology of network encodes conditional independence assertions: Considering Weather, Toothache, Cavity, and Catch, Weather is independent of the other variables. [Q] What else does not get any influence from others? Cavity influences Toothache and Catch. Toothache and Catch are conditionally independent given Cavity. 2.

Bayesian Networks9 Example - Burglary I'm at work, neighbor John calls to say my alarm is ringing, but neighbor Mary doesn't call. Sometimes it's set off by minor earthquakes. Is there a burglar? [Q] Random variables? Burglary, Earthquake, Alarm, JohnCalls, MaryCalls Network topology reflects "causal" or “influence” knowledge. [Q] In what way? A burglar can set the alarm off. An earthquake can set the alarm off. The alarm can cause Mary to call. The alarm can cause John to call. [Q] Can you draw the network then?

Bayesian Networks10 What is P(JohnCalls=T)? P(j) = P(j  a) + ??? = ??? + ??? = ??? What is P(Alarm=T)? P(a) = P(a  b  e) + P(a  b  e) + P(a  b  e) + P(a  b  e) = P(a|b  e) P(b  e) + P(a|b  e) P(b  e) + P(a|  b  e) P(  b  e) + P(a|  b  e) P(  b  e) = ??? John calls to say my alarm is ringing. P(a | b   e) [Q] What does it mean?

Bayesian Networks11 Compactness A CPT for Boolean X i with k Boolean parents has 2 k rows for the combinations of parent values. Each row requires one number p for X i = true (the number for X i = false is just 1-p) If each variable has no more than k parents, the complete network requires O(n · 2 k ) numbers I.e., grows linearly with n, vs. O(2 n ) for the full joint distribution But, for burglary net, = 10 numbers (vs = 31) 2.

Bayesian Networks12 How to answer queries in Burglary net The full joint distribution is defined as the product of the local conditional distributions: P(X 1, …, X n ) = π i = 1 P(X i | Parents(X i )) [Q] P(j  m  a   b   e) = ??? = P(j | a) P(m | a) P(a |  b,  e) P(  b) P(  e) =.9 *.7 *.001 *.999 *.998 ≈ n

Bayesian Networks13 I'm at work, neighbor John calls to say my alarm is ringing, but neighbor Mary doesn't call. Sometimes it's set off by minor earthquakes. What are the chances that the alarm was ringing and there is a burglar? P(j   m  a  b) = ??? [Q] How to deal with Earthquake? = P(j   m  a  b  e) + P(j   m  a  b   e) = P(j | a) P(  m | a) P(a | (b  e)) P(b) P(e) + P(j | a) P(  m | a) (P(a | (b  e)) P(b) P(  e) = P(j | a) (1 - P(m | a)) P(a | (b  e)) P(b) P(e) + P(j | a) P(  m | a) (P(a | (b  e)) P(b) P(  e) = ??? Mary calls to say alarm is ringing. what are the changes there is an earthquake? P(X 1, …, X n ) =  P(X i | Parents(X i ))

Bayesian Networks14 P(a) = ??? = P(a  b  e) + P(a  b   e) + P(a   b  e) + P(a   b   e) = P(a | (b  e)) P(b) P(e) + P(a | (b  e)) P(b) P(  e) + P(a | (  b  e)) P(  b) P(e) + P(a | (  b  e)) P(  b) P(  e)) = … P(j) = ??? P(X 1, …, X n ) =  P(X i | Parents(X i )) 2.

Bayesian Networks15 How to construct Bayesian networks 1. Choose an ordering of random variables X 1, …, X n 2. For i = 1 to n add X i to the network select parents from X 1, …, X i-1 such that P (X i | Parents(X i )) = P (X i | X 1,..., X i-1 ) i.e., Parents(X i ) are conditionally independent of the others. This choice of parents guarantees: P (X 1, …, X n ) = π i =1 P (X i | X 1, …, X i-1 ) (by chain rule) = π i =1 P (X i | Parents(X i )) (by construction) [Q] Which random variables first? n n

Bayesian Networks16 Suppose we choose the ordering M, J, A, B, E Example of a wrong order

Bayesian Networks17 Suppose we choose the ordering M, J, A, B, E P(J | M) = P(J)?

Bayesian Networks18 Suppose we choose the ordering M, J, A, B, E P(J | M) = P(J)? No P(A | J, M) = P(A | J)? P(A | J, M) = P(A)?

Bayesian Networks19 Suppose we choose the ordering M, J, A, B, E P(J | M) = P(J)? No P(A | J, M) = P(A | J)? P(A | J, M) = P(A)? No P(B | A, J, M) = P(B | A)? P(B | A, J, M) = P(B)?

Bayesian Networks20 Suppose we choose the ordering M, J, A, B, E P(J | M) = P(J)? No P(A | J, M) = P(A | J)? P(A | J, M) = P(A)? No P(B | A, J, M) = P(B | A)? Yes P(B | A, J, M) = P(B)? No P(E | B, A, J, M) = P(E | A)? P(E | B, A, J, M) = P(E | A, B)?

Bayesian Networks21 Suppose we choose the ordering M, J, A, B, E P(J | M) = P(J)? No P(A | J, M) = P(A | J)? P(A | J, M) = P(A)? No P(B | A, J, M) = P(B | A)? Yes P(B | A, J, M) = P(B)? No P(E | B, A, J, M) = P(E | A)? No P(E | B, A, J, M) = P(E | A, B)? Yes

Bayesian Networks22 Deciding conditional independence is hard in noncausal directions. (Causal models and conditional independence seem hardwired for humans!) Accessing conditional probabilities is hard in noncasual directions Network is less compact: = 13 numbers needed

Bayesian Networks23 How to choose a correct order? “Root causes” first, [Q] What can be a root? then the variables they directly influence, and so on, until the leaves Can we use causal relations? For each node, find child nodes (i.e., the child nodes caused by the node); Example of Weather, Toothache, Catch, and Cavity Topics 2.

Bayesian Networks24 Summary Bayesian networks provide a natural representation for (causally induced) conditional independence. Topology + CPTs = compact representation of joint distribution Generally easy for domain experts to construct Topics