A. Darwiche Sensitivity Analysis in Bayesian Networks Adnan Darwiche Computer Science Department

Slides:



Advertisements
Similar presentations
Analysis of Algorithms
Advertisements

CS188: Computational Models of Human Behavior
A. Darwiche Knowledge Compilation Jinbo Huang NICTA and ANU Slides made by Adnan Darwiche and Jinbo Huang.
Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Representations for KBS: Uncertainty & Decision Support
Bayesian Networks CSE 473. © Daniel S. Weld 2 Last Time Basic notions Atomic events Probabilities Joint distribution Inference by enumeration Independence.
A. Darwiche Bayesian Networks. A. Darwiche Reasoning Systems Diagnostics: Which component failed? Information retrieval: What document to retrieve? On-line.
1 Some Comments on Sebastiani et al Nature Genetics 37(4)2005.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 27, 2012.
A. Darwiche Sensitivity Analysis in Bayesian Networks Adnan Darwiche Computer Science Department
1 Test-Cost Sensitive Naïve Bayes Classification X. Chai, L. Deng, Q. Yang Dept. of Computer Science The Hong Kong University of Science and Technology.
Machine Learning in Practice Lecture 7 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
BAYESIAN NETWORKS CHAPTER#4 Book: Modeling and Reasoning with Bayesian Networks Author : Adnan Darwiche Publisher: CambridgeUniversity Press 2009.
An Introduction to Variational Methods for Graphical Models.
A Probabilistic Framework for Information Integration and Retrieval on the Semantic Web by Livia Predoiu, Heiner Stuckenschmidt Institute of Computer Science,
Assuming normally distributed data! Naïve Bayes Classifier.
A Differential Approach to Inference in Bayesian Networks - Adnan Darwiche Jiangbo Dang and Yimin Huang CSCE582 Bayesian Networks and Decision Graph.
Probabilistic Reasoning Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 14 (14.1, 14.2, 14.3, 14.4) Capturing uncertain knowledge Probabilistic.
Bayesian Belief Networks
Representing Uncertainty CSE 473. © Daniel S. Weld 2 Many Techniques Developed Fuzzy Logic Certainty Factors Non-monotonic logic Probability Only one.
1 Department of Computer Science and Engineering, University of South Carolina Issues for Discussion and Work Jan 2007  Choose meeting time.
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
1er. Escuela Red ProTIC - Tandil, de Abril, Bayesian Learning 5.1 Introduction –Bayesian learning algorithms calculate explicit probabilities.
. Approximate Inference Slides by Nir Friedman. When can we hope to approximate? Two situations: u Highly stochastic distributions “Far” evidence is discarded.
CPSC 422, Lecture 18Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18 Feb, 25, 2015 Slide Sources Raymond J. Mooney University of.
Rutgers CS440, Fall 2003 Introduction to Statistical Learning Reading: Ch. 20, Sec. 1-4, AIMA 2 nd Ed.
A Differential Approach to Inference in Bayesian Networks - Adnan Darwiche Jiangbo Dang and Yimin Huang CSCE582 Bayesian Networks and Decision Graphs.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
Jeff Howbert Introduction to Machine Learning Winter Classification Bayesian Classifiers.
Context-specific independence Graphical Models – Carlos Guestrin Carnegie Mellon University October 16 th, 2006 Readings: K&F: 4.1, 4.2, 4.3, 4.4,
Using Bayesian Networks to Analyze Expression Data N. Friedman, M. Linial, I. Nachman, D. Hebrew University.
Bayesian Learning By Porchelvi Vijayakumar. Cognitive Science Current Problem: How do children learn and how do they get it right?
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
Bayesian networks. Motivation We saw that the full joint probability can be used to answer any question about the domain, but can become intractable as.
Introduction to Bayesian Networks
Classification Techniques: Bayesian Classification
Inference Complexity As Learning Bias Daniel Lowd Dept. of Computer and Information Science University of Oregon Joint work with Pedro Domingos.
Announcements Project 4: Ghostbusters Homework 7
Slides for “Data Mining” by I. H. Witten and E. Frank.
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
CHAPTER 6 Naive Bayes Models for Classification. QUESTION????
Wei Sun and KC Chang George Mason University March 2008 Convergence Study of Message Passing In Arbitrary Continuous Bayesian.
29 August 2013 Venkat Naïve Bayesian on CDF Pair Scores.
Random Geometric Graph Model Model for ad hoc/sensor networks n nodes placed in d-dim space Connectivity threshold r Two nodes u,v connected iff ||u-v||
1 Machine Learning: Lecture 6 Bayesian Learning (Based on Chapter 6 of Mitchell T.., Machine Learning, 1997)
CSE 473 Uncertainty. © UW CSE AI Faculty 2 Many Techniques Developed Fuzzy Logic Certainty Factors Non-monotonic logic Probability Only one has stood.
CHAPTER 3: BAYESIAN DECISION THEORY. Making Decision Under Uncertainty Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Conditional Independence As with absolute independence, the equivalent forms of X and Y being conditionally independent given Z can also be used: P(X|Y,
Probabilistic Reasoning Inference and Relational Bayesian Networks.
Naive Bayes Classifier. REVIEW: Bayesian Methods Our focus this lecture: – Learning and classification methods based on probability theory. Bayes theorem.
CS 541: Artificial Intelligence Lecture VII: Inference in Bayesian Networks.
1 1)Bayes’ Theorem 2)MAP, ML Hypothesis 3)Bayes optimal & Naïve Bayes classifiers IES 511 Machine Learning Dr. Türker İnce (Lecture notes by Prof. T. M.
Compiling Bayesian Networks Using Variable Elimination
Instructor: Eyal Amir Grad TAs: Wen Pu, Yonatan Bisk
CS 541: Artificial Intelligence
Reasoning Under Uncertainty: More on BNets structure and construction
Boosted Augmented Naive Bayes. Efficient discriminative learning of
Inference in Bayesian Networks
Solving MAP Exactly by Searching on Compiled Arithmetic Circuits
A New Algorithm for Computing Upper Bounds for Functional EmajSAT
Computer Science Department
Reasoning Under Uncertainty: More on BNets structure and construction
Encoding CNFs to Enhance Component Analysis
Binary Decision Diagrams
Pattern Recognition and Image Analysis
CAP 5636 – Advanced Artificial Intelligence
CS 188: Artificial Intelligence
Bayesian Learning Chapter
Pegna, J.M., Lozano, J.A., and Larragnaga, P.
Context-specific independence
Presentation transcript:

A. Darwiche Sensitivity Analysis in Bayesian Networks Adnan Darwiche Computer Science Department

A. Darwiche Bayesian network classifiers Given: –A Bayesian network N –A class variable C –A set of variables, attributes E = {E 1, …, E n } Each instantiation e is called an instance –A probability threshold p Define Bayesian network classifier F:

A. Darwiche Naïve Bayes Classifiers PPr(p) yes0.87 no0.13 PUPr(u|p) yes-ve0.27 no+ve0.107 PBPr(b|p) yes-ve0.36 no+ve0.106 PSPr(s|p) yes-ve0.10 no+ve0.01 UBSPr(P = yes | u, b, s) +ve ve -ve ve-ve+ve ve-ve ve+ve ve+ve-ve ve +ve ve  0.9? Yes No Yes No Yes Which sets of test results confirm pregnancy, with probability no less than 90%? Pregnant? (P) Urine test (U) Blood test (B) Scanning test (S)

A. Darwiche Reasoning about Bayesian network classifiers Given N and N’, do they induce the same classifier? If not, which, and how many, instances do they disagree on? Given N, what are the allowable changes to a CPT which will not change the current classifier? How many distinct classifiers can be induced by changing some CPT?

A. Darwiche Reasoning about Bayesian network classifiers We can answer these questions by enumerating all instances e explicitly However, this is often infeasible given the exponential number of instances Instead, we propose to build a tractable logical representation of the classifier F N This allows us to answer the above questions in time linear in the size of the representation

A. Darwiche From Numbers to Decisions + Probabilistic Inference 0.87yes 0.13no Pr(p)P 0.27-veyes no P ve Pr(u|p)U 0.36-veyes no P ve Pr(b|p)B 0.10-veyes no P 0.01+ve Pr(s|p)S Pregnant? (P) Urine test (U) Blood test (B) Scanning test (S) Decision Function Test results: U, B, S Yes, No

A. Darwiche U +ve -ve B S Yes +ve -ve No -ve +ve Situation: U=+ve, B=-ve, S=-ve 0.87yes 0.13no Pr(p)P 0.27-veyes no P ve Pr(u|p)U 0.36-veyes no P ve Pr(b|p)B 0.10-veyes no P 0.01+ve Pr(s|p)S Pregnant? (P) Urine test (U) Blood test (B) Scanning test (S) Ordered Decision Diagram + Probabilistic Inference From Numbers to Decisions

A. Darwiche X1 X2 X3 1 0 Binary Decision Diagram Test-once property

A. Darwiche Improving Reliability of Sensors Currently False negative 27.0% False positive 10.7% Pregnant? (P) Urine test (U) Blood test (B) Scanning test (S) Same decisions (in all situations) if new test is: False negative 10% False positive 5% Different decisions (in some situations) if new test: False negative 5% False positive 2.5% Can characterize these situations, compute their likelihood, analyze their properties Yes if > 90%

A. Darwiche Adding New Sensors Pregnant? (P) Urine test (U) Blood test (B) Scanning test (S) New test (N) Can characterize these situations, compute their likelihood, analyze their properties Same decisions (in all situations) if: False negative 40% False positive 20% Different (in some situations) decisions if: False negative 20% False positive 10% Yes if > 90%

A. Darwiche Naïve Bayes classifier Class variable C Attributes E PPr(p) yes0.87 no0.13 PUPr(u|p) yes-ve0.27 no+ve0.107 PBPr(b|p) yes-ve0.36 no+ve0.106 PSPr(s|p) yes-ve0.10 no+ve0.01

A. Darwiche Naïve Bayes classifier Prior log-oddsWeight of evidence e i

A. Darwiche Changing the prior log-odds in a naïve Bayes classifier If we change the CPT of C, thereby changing the prior log-odds from log O(c) to log O’(c), will we still have the same classifier?

A. Darwiche Changing the prior log-odds in a naïve Bayes classifier 

A. Darwiche Equivalence of NB classifiers

A. Darwiche Equivalence of NB classifiers Equivalent iff prior of P in F N’  [0.684, 0.970) Change prior of P

A. Darwiche Path 1Path 2 Sub-ODD D 1 Sub-ODD D 2

A. Darwiche Path 1Path 2 Sub-ODD D 1 = D 2

A. Darwiche Theoretical results of algorithm Space complexity: –Total number of nodes in the ODD  O(b n/2 ) Time complexity: –O(nb n/2 ) Improves greatly over brute-force approach: –Total number of instances = O(b n )

A. Darwiche Experimental results of algorithm Network# Attributes# Instances# Nodes bound# Nodes Tic-tac-toe Votes Spect22 4  Breast-cancer-w9 1  Hepatitis19 2  Kr-vs-kp36 1  Mushroom22 1  

A. Darwiche

A. Darwiche