Bayesian Nets and Applications Next class: machine learning C. 18.1, 18.2 Homework due next class Questions on the homework? Prof. McKeown will not hold.

Slides:



Advertisements
Similar presentations
Bayesian networks Chapter 14 Section 1 – 2. Outline Syntax Semantics Exact computation.
Advertisements

Markov Networks Alan Ritter.
Identifying Conditional Independencies in Bayes Nets Lecture 4.
Bayesian Networks Chapter 14 Section 1, 2, 4. Bayesian networks A simple, graphical notation for conditional independence assertions and hence for compact.
CPSC 322, Lecture 26Slide 1 Reasoning Under Uncertainty: Belief Networks Computer Science cpsc322, Lecture 27 (Textbook Chpt 6.3) March, 16, 2009.
Toothache  toothache catch  catch catch  catch cavity  cavity Joint PDF.
Review: Bayesian learning and inference
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11
A Differential Approach to Inference in Bayesian Networks - Adnan Darwiche Jiangbo Dang and Yimin Huang CSCE582 Bayesian Networks and Decision Graph.
Probabilistic Reasoning Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 14 (14.1, 14.2, 14.3, 14.4) Capturing uncertain knowledge Probabilistic.
Bayesian networks Chapter 14 Section 1 – 2.
Bayesian Belief Networks
Bayesian Nets and Applications Today’s Reading: C. 14 Next class: machine learning C. 18.1, 18.2 Questions on the homework?
Part 2 of 3: Bayesian Network and Dynamic Bayesian Network.
Example applications of Bayesian networks
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
Bayesian Networks Alan Ritter.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
Bayesian networks More commonly called graphical models A way to depict conditional independence relationships between random variables A compact specification.
Judgment and Decision Making in Information Systems Computing with Influence Diagrams and the PathFinder Project Yuval Shahar, M.D., Ph.D.
Quiz 4: Mean: 7.0/8.0 (= 88%) Median: 7.5/8.0 (= 94%)
Bayes’ Nets  A Bayes’ net is an efficient encoding of a probabilistic model of a domain  Questions we can ask:  Inference: given a fixed BN, what is.
Bayesian networks Chapter 14. Outline Syntax Semantics.
Bayesian networks Chapter 14 Section 1 – 2. Bayesian networks A simple, graphical notation for conditional independence assertions and hence for compact.
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
Bayesian networks. Motivation We saw that the full joint probability can be used to answer any question about the domain, but can become intractable as.
1 Chapter 14 Probabilistic Reasoning. 2 Outline Syntax of Bayesian networks Semantics of Bayesian networks Efficient representation of conditional distributions.
Bayesian Networks for Data Mining David Heckerman Microsoft Research (Data Mining and Knowledge Discovery 1, (1997))
Probabilistic Reasoning ECE457 Applied Artificial Intelligence Spring 2007 Lecture #9.
1 Monte Carlo Artificial Intelligence: Bayesian Networks.
Introduction to Bayesian Networks
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
Bayesian Nets and Applications. Naïve Bayes 2  What happens if we have more than one piece of evidence?  If we can assume conditional independence 
Announcements Project 4: Ghostbusters Homework 7
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS.
Probabilistic Reasoning [Ch. 14] Bayes Networks – Part 1 ◦Syntax ◦Semantics ◦Parameterized distributions Inference – Part2 ◦Exact inference by enumeration.
Marginalization & Conditioning Marginalization (summing out): for any sets of variables Y and Z: Conditioning(variant of marginalization):
Review: Bayesian inference  A general scenario:  Query variables: X  Evidence (observed) variables and their values: E = e  Unobserved variables: Y.
1 Tournament Not complete Processing will begin again tonight, 7:30PM until wee hours Friday, 8-5. Extra Credit 5 points for passing screening, in tournament.
1 Bayesian Networks: A Tutorial. 2 Introduction Suppose you are trying to determine if a patient has tuberculosis. You observe the following symptoms:
Quiz 3: Mean: 9.2 Median: 9.75 Go over problem 1.
Introduction on Graphic Models
Belief Networks Kostas Kontogiannis E&CE 457. Belief Networks A belief network is a graph in which the following holds: –A set of random variables makes.
CPSC 322, Lecture 26Slide 1 Reasoning Under Uncertainty: Belief Networks Computer Science cpsc322, Lecture 27 (Textbook Chpt 6.3) Nov, 13, 2013.
Weng-Keen Wong, Oregon State University © Bayesian Networks: A Tutorial Weng-Keen Wong School of Electrical Engineering and Computer Science Oregon.
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
Artificial Intelligence Bayes’ Nets: Independence Instructors: David Suter and Qince Li Course Harbin Institute of Technology [Many slides.
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11 CS479/679 Pattern Recognition Dr. George Bebis.
CS 2750: Machine Learning Directed Graphical Models
Reasoning Under Uncertainty: More on BNets structure and construction
Bayesian networks Chapter 14 Section 1 – 2.
Presented By S.Yamuna AP/CSE
Read R&N Ch Next lecture: Read R&N
Reasoning Under Uncertainty: More on BNets structure and construction
Read R&N Ch Next lecture: Read R&N
Structure and Semantics of BN
Pattern Recognition and Image Analysis
CAP 5636 – Advanced Artificial Intelligence
CS 188: Artificial Intelligence
Class #16 – Tuesday, October 26
CS 188: Artificial Intelligence Spring 2007
Structure and Semantics of BN
Bayesian networks Chapter 14 Section 1 – 2.
Probabilistic Reasoning
CS 188: Artificial Intelligence Spring 2006
Read R&N Ch Next lecture: Read R&N
Chapter 14 February 26, 2004.
CS 188: Artificial Intelligence Fall 2008
Presentation transcript:

Bayesian Nets and Applications Next class: machine learning C. 18.1, 18.2 Homework due next class Questions on the homework? Prof. McKeown will not hold office hours today

2 Bayesian Networks A directed acyclic graph in which each node is annotated with quantitative probability information –A set of random variables makes up the network nodes –A set of directed links connects pairs of nodes. If there is an arrow from node X to node Y, X is a parent of Y –Each node X i has a conditional probability distributionP(X i |Parents(X i ) that quantifies the effect of the parents on the node

3 Example Topology of network encodes conditional independence assumptions

4 Smart Good test taker Understands material Hard working Exam GradeHomework Grade

5 Smart Good test taker Understands material Hard working Exam GradeHomework Grade Smart TrueFalse.5 Hard Working TrueFalse.7.3 SGood Test Taker TrueFalse True False SHWUM TrueFalse True TrueFalse.6.4 FalseTrue.6.4 False.2.8

6 Conditional Probability Tables Smart TrueFalse.5 Hard Working TrueFalse.7.3 SGood Test Taker TrueFalse True False SHWUM TrueFalse True TrueFalse.6.4 FalseTrue.6.4 False.2.8 GTTUMExam Grade ABCDF True TrueFalse FalseTrue False Homework Grade UMABCDF True False

7 Compactness A CPT for Boolean X i with k Boolean parents has 2 k rows for the combinations of parent values Each row requires one number p for X i =true (the number for X i =false is just 1-p) If each variable has no more than k parents, the complete network requires O(nx2 k ) numbers Grows linearly with n vs O(2 n ) for the full joint distribution Student net: =11 numbers (vs. 26-1)=31

8 Global Semantics/Evaluation Global semantics defines the full joint distribution as the product of the local conditional distributions: P(x 1,…,x n )=∏ i n =1 P(x i | Parents(X i )) e.g., P(EG=AΛGTΛ⌐UMΛSΛHW)

9 Global Semantics Global semantics defines the full joint distribution as the product of the local conditional distributions: P(X 1,…,X n )=∏ i n =1 P(X i |Parents(X i )) e.g., Observations:S, HW, not UM, will I get an A? P(EG=AΛGTΛ⌐UMΛSΛHW) = P(EG=A|GT Λ⌐UM)*P(GT|S)*P(⌐UM |HW ΛS)*P(S)*P(HW)

10 Conditional Independence and Network Structure The graphical structure of a Bayesian network forces certain conditional independences to hold regardless of the CPTs. This can be determined by the d- separation criteria

11 a b c a b c b a c Linear Converging Diverging

12 D-separation (opposite of d- connecting) A path from q to r is d-connecting with respect to the evidence nodes E if every interior node n in the path has the property that either It is linear or diverging and is not a member of E It is converging and either n or one of its decendants is in E If a path is not d-connecting (is d-separated), the nodes are conditionally independent given E

13 Smart Good test taker Understands material Hard working Exam GradeHomework Grade

14 S and EG are not independent given GTT S and HG are independent given UM

Medical Application of Bayesian Networks: Pathfinder

16 Pathfinder Domain: hematopathology diagnosis –Microscopic interpretation of lymph-node biopsies –Given: 100s of histologic features appearing in lymph node sections –Goal: identify disease type malignant or benign –Difficult for physicians

17 Pathfinder System Bayesian Net implementation Reasons about 60 malignant and benign diseases of the lymph node Considers evidence about status of up to 100 morphological features presenting in lymph node tissue Contains 105,000 subjectively-derived probabilities

18

19 Commercialization Intellipath Integrates with videodisc libraries of histopathology slides Pathologists working with the system make significantly more correct diagnoses than those working without Several hundred commercial systems in place worldwide

20 Sequential Diagnosis

21 Features Structured into a set of 2-10 mutually exclusive values Pseudofollicularity Absent, slight, moderate, prominent Represent evidence provided by a feature as F 1,F 2, … F n

22 Value of information User enters findings from microscopic analysis of tissue Probabilistic reasoner assigns level of belief to different diagnoses Value of information determines which tests to perform next Full disease utility model making use of life and death decision making Cost of tests Cost of misdiagnoses

23

24

25 Group Discrimination Strategy Select questions based on their ability to discriminate between disease classes For given differential diagnoisis, select most specific level of hierarchy and selects questions to discriminate among groups Less efficient Larger number of questions asked

26

27

28 Other Bayesian Net Applications Lumiere – Who knows what it is?

29 Other Bayesian Net Applications Lumiere Single most widely distributed application of BN Microsoft Office Assistant Infer a user’s goals and needs using evidence about user background, actions and queries VISTA Help NASA engineers in round-the-clock monitoring of each of the Space Shuttle’s orbiters subsystem Time critical, high impact Interpret telemetry and provide advice about likely failures Direct engineers to the best information In use for several years Microsoft Pregnancy and Child Care What questions to ask next to diagnose illness of a child

30 Other Bayesian Net Applications Speech Recognition Text Summarization Language processing tasks in general