Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?

Slides:



Advertisements
Similar presentations
Bayesian networks Chapter 14 Section 1 – 2. Outline Syntax Semantics Exact computation.
Advertisements

Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Bayesian Networks CSE 473. © Daniel S. Weld 2 Last Time Basic notions Atomic events Probabilities Joint distribution Inference by enumeration Independence.
BAYESIAN NETWORKS. Bayesian Network Motivation  We want a representation and reasoning system that is based on conditional independence  Compact yet.
Exact Inference in Bayes Nets
Uncertainty Everyday reasoning and decision making is based on uncertain evidence and inferences. Classical logic only allows conclusions to be strictly.
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11
10/24  Exam on 10/26 (Lei Tang and Will Cushing to proctor)
Probabilistic Reasoning Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 14 (14.1, 14.2, 14.3, 14.4) Capturing uncertain knowledge Probabilistic.
Bayesian networks Chapter 14 Section 1 – 2.
Bayesian Belief Networks
Bayesian Networks. Graphical Models Bayesian networks Conditional random fields etc.
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
Representing Uncertainty CSE 473. © Daniel S. Weld 2 Many Techniques Developed Fuzzy Logic Certainty Factors Non-monotonic logic Probability Only one.
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
CS 188: Artificial Intelligence Spring 2007 Lecture 14: Bayes Nets III 3/1/2007 Srini Narayanan – ICSI and UC Berkeley.
CS 188: Artificial Intelligence Fall 2006 Lecture 17: Bayes Nets III 10/26/2006 Dan Klein – UC Berkeley.
10/22  Homework 3 returned; solutions posted  Homework 4 socket opened  Project 3 assigned  Mid-term on Wednesday  (Optional) Review session Tuesday.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
Quiz 4: Mean: 7.0/8.0 (= 88%) Median: 7.5/8.0 (= 94%)
Read R&N Ch Next lecture: Read R&N
Bayesian networks Chapter 14. Outline Syntax Semantics.
Bayesian Belief Networks. What does it mean for two variables to be independent? Consider a multidimensional distribution p(x). If for two features we.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
Bayesian networks. Motivation We saw that the full joint probability can be used to answer any question about the domain, but can become intractable as.
1 Chapter 14 Probabilistic Reasoning. 2 Outline Syntax of Bayesian networks Semantics of Bayesian networks Efficient representation of conditional distributions.
2 Syntax of Bayesian networks Semantics of Bayesian networks Efficient representation of conditional distributions Exact inference by enumeration Exact.
Bayesian Statistics and Belief Networks. Overview Book: Ch 13,14 Refresher on Probability Bayesian classifiers Belief Networks / Bayesian Networks.
1 Monte Carlo Artificial Intelligence: Bayesian Networks.
Introduction to Bayesian Networks
1 CMSC 671 Fall 2001 Class #21 – Tuesday, November 13.
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
Marginalization & Conditioning Marginalization (summing out): for any sets of variables Y and Z: Conditioning(variant of marginalization):
CHAPTER 5 Probability Theory (continued) Introduction to Bayesian Networks.
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Bayesian networks and their application in circuit reliability estimation Erin Taylor.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
Inference Algorithms for Bayes Networks
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
Conditional Independence As with absolute independence, the equivalent forms of X and Y being conditionally independent given Z can also be used: P(X|Y,
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016.
Web-Mining Agents Data Mining Prof. Dr. Ralf Möller Universität zu Lübeck Institut für Informationssysteme Karsten Martiny (Übungen)
CS 541: Artificial Intelligence Lecture VII: Inference in Bayesian Networks.
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11 CS479/679 Pattern Recognition Dr. George Bebis.
CS 2750: Machine Learning Directed Graphical Models
Bayesian networks Chapter 14 Section 1 – 2.
Presented By S.Yamuna AP/CSE
Qian Liu CSE spring University of Pennsylvania
Inference in Bayesian Networks
CS b553: Algorithms for Optimization and Learning
Read R&N Ch Next lecture: Read R&N
CS 4/527: Artificial Intelligence
CSCI 121 Special Topics: Bayesian Networks Lecture #2: Bayes Nets
Read R&N Ch Next lecture: Read R&N
Representing Uncertainty
Structure and Semantics of BN
CAP 5636 – Advanced Artificial Intelligence
Bayesian Statistics and Belief Networks
CS 188: Artificial Intelligence
Class #19 – Tuesday, November 3
CS 188: Artificial Intelligence
CS 188: Artificial Intelligence Fall 2008
Class #16 – Tuesday, October 26
Structure and Semantics of BN
Belief Networks CS121 – Winter 2003 Belief Networks.
Bayesian networks Chapter 14 Section 1 – 2.
Probabilistic Reasoning
Presentation transcript:

Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?

Issues Representational Power –allows for unknown, uncertain information Inference –Question: What is Probability of X if E is true. –Processing: in general, exponential Acquisition or Learning –network: human input – probabilities: data+ learning

Bayesian Network Directed Acyclic Graph Nodes are RV’s Edges denote dependencies Root nodes = nodes without predecessors –prior probability table Non-root nodes –conditional probabilites for all predecessors

Pearl Alarm Example B -> A E ->A A->JC A->MC P(B) =.001 P(-B) =.999 P(E) =.002 P(-E) =.998 etc. P(A|B&E) =.95 P(A|B&-E) =.94 P(A|-B&E) =.29 P(A|-B&-E) =.001 P(JC|A) =.90 P(JC|-A) =.05 P(MC|A) =.70 P(MC|-A) =.01

Joint Probability yields all Event = fully specified values for RVs. Prob of event: P(x1,x2,..xn) = P(x1|Parents(X1)*..P(xn|Parents(Xn)) E.g. P(j&m&a&-b&-e) = P(j|a)*P(m|a)*P(a|-b^-e)*P(-b)*P(-e) =.9*.7*.001*.999*..998 = Do this for all events and then sum as needed. Yield exact probability

Summing out P(b) from full joint = sum over all events with b = true (marginalization) –Silly: must be.001. P(MC|JC) = linked by alarm, sum still too much but not so apparent Need to Answer: what RVs are independent of others depending on the evidence. We’re skipping: Market Blankets

Probability Calculation Cost For example, P( 5 boolean variables) requires 2^5 entries. In general 2^n. For Bayes net, only need tables for all conditional probabilities and priors. If max k inputs to a node, and n RVs, then need n*2^k table entries. Computation can be reduced, but difficult.

Approximate Inference Simple Sampling Use BayesNetwork as a generative model Eg. generate examples, via topological order. Generates examples with appropriate distribution. Now use examples to estimate probabilities.

Sampling -> probabilities Generate examples with proper probability density. Use the ordering of the nodes to construct events. Finally use counting to yield an estimate of the exact probability.

Estimate P(JC|MC) 1.Do a large number of times 1.Use prior tables to compute “root events” 2.Say b = f and e = g 3.Use conditional tables to compute internal nodes values (IN ORDER) 4.Say a = false 5.Say JC = false and MC = true 2.Count the number of appropriate events 3.Note many entries irrelevant: In this case only if MC = true is event considered. Markov Chain Monte Carlo will only construct appropriate events.

Confidence of Estimate Given n examples and k are heads. How many examples needed to be 99% certain that k/n is within.01 of the true p. From statistic: Mean = np, Variance = npq For confidence of.99, t = 3.25 (table) 3.25*sqrt(pq/N) N >6,400. Correct probabilities not needed: just correct ordering.

Applications Bayesian Networks extended to decision theory. –decision nodes have actions attached –value nodes indicate expect utility Pathfinder (heckerman): medical diagnosis –adds utility theory (decision theory) –some actions specific tests –60 disease, 130 features Research Arena